Ceph使用---Crush Map进阶

一、Ceph Crush Map介绍

ceph 集群中由 mon 服务器维护的的五种运行图:

Monitor map #监视器运行图
OSD map #OSD 运行图
PG map #PG 运行图
Crush map #(Controllers replication under scalable hashing #可控的、 可复制的、 可伸缩的一致性 hash 算法。
MDS map #cephfs metadata 运行图
crush 运行图,当新建存储池时会基于 OSD map 创建出新的 PG 组合列表用于存储数据

crush 算法针对目的节点的选择:

目前有 5 种算法来实现节点的选择, 包括 Uniform、 List、 Tree、 Straw、 Straw2, 早期版本使用的是 ceph 项目的发起者发明的算法 straw, 目前已经发展社区优化的 straw2 版本。

straw(抽签算法)

抽签是指挑取一个最长的签, 而这个签值就是 OSD 的权重, 当创建存储池的时候会向 PG分配 OSD, straw 算法会遍历当前可用的 OSD 并优先使用中签的 OSD, 以让权重高的 OSD被分配较多的 PG 以存储更多
的数据。

二、PG 与 OSD 映射调整

默认情况下,crush 算法自行对创建的 pool 中的 PG 分配 OSD, 但是可以手动基于权重设置crush 算法分配数据的倾向性, 比如 1T 的磁盘权重是 1, 2T 的就是 2, 推荐使用相同大小的设备。

查看当前状态:

weight 表示设备(device)的容量相对值, 比如 1TB 对应 1.00, 那么 500G 的 OSD 的 weight就应该是 0.5, weight 是基于磁盘空间分配 PG 的数量, 让 crush 算法尽可能往磁盘空间大的 OSD 多分
配 PG.往磁盘空间小的 OSD 分配较少的 PG。 Reweight 参数的目的是重新平衡 ceph 的 CRUSH 算法随机分配的 PG, 默认的分配是概率上的均衡, 即使 OSD 都是一样的磁盘空间也会产生一些 PG 分布不均匀的情况, 此时可以通过调整 reweight 参数,
让 ceph 集群立即重新平衡当前磁盘的 PG, 以达到数据均衡分布的目的, REWEIGHT 是 PG 已经分配完成, 要在 ceph 集群重新平衡 PG 的分布。

修改 WEIGHT 并验证:

[root@ceph-deploy ~]# ceph osd crush reweight osd.1 1.5 #此操作属于高危操作,会立即触发ceph集群数据重平衡,谨慎操作!!!!
reweighted item id 1 name 'osd.1' to 1.5 in crush map
[root@ceph-deploy ~]# ceph osd df 
ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS
 0    hdd  0.09769   1.00000  100 GiB  770 MiB  719 MiB   20 KiB   51 MiB   99 GiB  0.75  1.14  133      up
 1    hdd  1.50000   1.00000  100 GiB  695 MiB  627 MiB    7 KiB   68 MiB   99 GiB  0.68  1.03  147      up
 2    hdd  0.09769   1.00000  100 GiB  555 MiB  511 MiB    6 KiB   44 MiB   99 GiB  0.54  0.82  118      up
 3    hdd  0.09769   1.00000  100 GiB  679 MiB  625 MiB   12 KiB   53 MiB   99 GiB  0.66  1.00  121      up
 4    hdd  0.09769   1.00000  100 GiB  652 MiB  595 MiB   13 KiB   58 MiB   99 GiB  0.64  0.97  132      up
 5    hdd  0.09769   1.00000  100 GiB  688 MiB  636 MiB    8 KiB   52 MiB   99 GiB  0.67  1.02  147      up
 6    hdd  0.09769   1.00000  100 GiB  638 MiB  571 MiB    7 KiB   67 MiB   99 GiB  0.62  0.94  125      up
 7    hdd  0.09769   1.00000  100 GiB  669 MiB  611 MiB   13 KiB   58 MiB   99 GiB  0.65  0.99  141      up
 8    hdd  0.09769   1.00000  100 GiB  734 MiB  674 MiB   13 KiB   60 MiB   99 GiB  0.72  1.09  135      up
                       TOTAL  900 GiB  5.9 GiB  5.4 GiB  104 KiB  510 MiB  894 GiB  0.66                   
MIN/MAX VAR: 0.82/1.14  STDDEV: 0.06
[root@ceph-deploy ~]# ceph -s
  cluster:
    id:     8dc32c41-121c-49df-9554-dfb7deb8c975
    health: HEALTH_WARN
            Degraded data redundancy: 1484/2529 objects degraded (58.679%), 151 pgs degraded  #此操作属于高危操作,会立即触发ceph集群数据重平衡,谨慎操作!!!!
 
  services:
    mon: 3 daemons, quorum ceph-mon1,ceph-mon2,ceph-mon3 (age 21h)
    mgr: ceph-mgr1(active, since 21h), standbys: ceph-mgr2
    mds: 2/2 daemons up, 2 standby
    osd: 9 osds: 9 up (since 21h), 9 in (since 6d); 58 remapped pgs
    rgw: 2 daemons active (2 hosts, 1 zones)
 
  data:
    volumes: 1/1 healthy
    pools:   12 pools, 401 pgs
    objects: 843 objects, 1.8 GiB
    usage:   6.0 GiB used, 894 GiB / 900 GiB avail
    pgs:     1484/2529 objects degraded (58.679%)
             571/2529 objects misplaced (22.578%)
             211 active+clean
             127 active+recovery_wait+degraded
             31  active+remapped+backfill_wait
             24  active+recovery_wait+undersized+degraded+remapped
             4   active+recovery_wait
             3   active+recovery_wait+remapped
             1   active+recovering
 
  io:
    client:   511 KiB/s wr, 0 op/s rd, 0 op/s wr
    recovery: 4.5 MiB/s, 1 keys/s, 3 objects/s  #数据量过大,此处值阈值不放开,会造成osd拥塞,最终集群无法正常提供服务
 
[root@ceph-deploy ~]# 

[root@ceph-deploy ~]# ceph health detail 

HEALTH_WARN Degraded data redundancy: 1143/2529 objects degraded (45.196%), 111 pgs degraded
[WRN] PG_DEGRADED: Degraded data redundancy: 1143/2529 objects degraded (45.196%), 111 pgs degraded
    pg 2.0 is active+recovery_wait+degraded, acting [3,1,8]
    pg 2.1 is active+recovery_wait+degraded, acting [1,3,8]
    pg 2.2 is active+recovery_wait+degraded, acting [1,5,6]
    pg 2.11 is active+recovery_wait+degraded, acting [1,3,8]
    pg 2.13 is active+recovery_wait+degraded, acting [1,4,6]
    pg 2.15 is active+recovery_wait+degraded, acting [1,8,5]
    pg 2.16 is active+recovery_wait+degraded, acting [1,7,4]
    pg 2.18 is active+recovery_wait+degraded, acting [1,6,3]
    pg 2.19 is active+recovery_wait+degraded, acting [1,7,3]
    pg 2.1a is active+recovery_wait+degraded, acting [1,4,7]
    pg 2.1b is active+recovery_wait+degraded, acting [6,1,5]
    pg 2.39 is active+recovery_wait+degraded, acting [7,5,1]
    pg 2.3d is active+recovery_wait+degraded, acting [1,7,4]
    pg 2.3e is active+recovery_wait+degraded, acting [7,1,4]
    pg 3.f is active+recovery_wait+degraded, acting [1,8,4]
    pg 6.5 is active+recovery_wait+degraded, acting [0,4,8]
    pg 6.6 is active+recovery_wait+degraded, acting [1,3,8]
    pg 6.7 is active+recovery_wait+degraded, acting [1,6,3]
    pg 8.4 is active+recovery_wait+degraded, acting [1,5,7]
    pg 8.5 is active+recovery_wait+degraded, acting [6,2,5]
    pg 8.6 is active+recovery_wait+degraded, acting [6,0,3]
    pg 8.7 is active+recovery_wait+degraded, acting [1,6,3]
    pg 8.8 is active+recovery_wait+degraded, acting [1,7,5]
    pg 8.9 is active+recovery_wait+degraded, acting [1,4,7]
    pg 8.a is active+recovery_wait+degraded, acting [1,7,3]
    pg 8.10 is active+recovery_wait+degraded, acting [3,7,1]
    pg 8.13 is active+recovery_wait+degraded, acting [1,7,5]
    pg 8.18 is active+recovery_wait+degraded, acting [1,5,6]
    pg 8.1a is active+recovery_wait+degraded, acting [1,3,7]
    pg 8.1b is active+recovery_wait+degraded, acting [1,5,8]
    pg 8.1c is active+recovery_wait+degraded, acting [1,3,7]
    pg 8.1d is active+recovery_wait+degraded, acting [1,4,8]
    pg 8.1f is active+recovery_wait+degraded, acting [7,1,3]
    pg 8.37 is active+recovery_wait+degraded, acting [1,3,6]
    pg 13.0 is active+recovery_wait+undersized+degraded+remapped, acting [7,4]
    pg 13.2 is active+recovery_wait+undersized+degraded+remapped, acting [5,6]
    pg 13.3 is active+recovery_wait+undersized+degraded+remapped, acting [0,4]
    pg 13.16 is active+recovery_wait+undersized+degraded+remapped, acting [8,3]
    pg 13.18 is active+recovery_wait+undersized+degraded+remapped, acting [5,8]
    pg 13.1c is active+recovery_wait+undersized+degraded+remapped, acting [5,1]
    pg 13.1d is active+recovery_wait+undersized+degraded+remapped, acting [3,7]
    pg 13.1e is active+recovery_wait+undersized+degraded+remapped, acting [1,5]
    pg 13.1f is active+recovery_wait+undersized+degraded+remapped, acting [0,5]
    pg 15.1 is active+recovery_wait+degraded, acting [1,3,7]
    pg 15.3 is active+recovery_wait+degraded, acting [1,6,4]
    pg 15.c is active+recovery_wait+degraded, acting [7,1,4]
    pg 15.e is active+recovery_wait+degraded, acting [1,6,4]
    pg 15.15 is active+recovery_wait+degraded, acting [5,7,1]
    pg 15.16 is active+recovery_wait+degraded, acting [1,4,8]
    pg 15.1d is active+recovery_wait+degraded, acting [1,7,3]
    pg 15.1f is active+recovery_wait+degraded, acting [1,4,8]

修改 REWEIGHT 并验证:
OSD 的 REWEIGHT 的值默认为 1, 值可以调整, 范围在 0~1 之间, 值越低 PG 越小, 如果调整了任何一个 OSD 的 REWEIGHT 值, 那么 OSD 的 PG 会立即和其它 OSD 进行重新平衡, 即数据的重新分配, 用于当某个 OSD 的 PG 相对较多需要降低其 PG 数量的场景。

[root@ceph-deploy ~]# ceph osd reweight 2 0.6
reweighted osd.2 to 0.6 (9999)
[root@ceph-deploy ~]#
[root@ceph-deploy ~]# ceph osd df
ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS
 0    hdd  0.09769   1.00000  100 GiB  828 MiB  771 MiB   20 KiB   57 MiB   99 GiB  0.81  1.22  163      up
 1    hdd  0.09769   1.00000  100 GiB  772 MiB  677 MiB    7 KiB   96 MiB   99 GiB  0.75  1.13  153      up
 2    hdd  0.09769   0.59999  100 GiB  454 MiB  400 MiB    6 KiB   53 MiB  100 GiB  0.44  0.67   85      up
 3    hdd  0.09769   1.00000  100 GiB  684 MiB  623 MiB   12 KiB   61 MiB   99 GiB  0.67  1.00  127      up
 4    hdd  0.09769   1.00000  100 GiB  661 MiB  592 MiB   13 KiB   69 MiB   99 GiB  0.65  0.97  135      up
 5    hdd  0.09769   1.00000  100 GiB  696 MiB  633 MiB    8 KiB   62 MiB   99 GiB  0.68  1.02  139      up
 6    hdd  0.09769   1.00000  100 GiB  662 MiB  580 MiB    7 KiB   82 MiB   99 GiB  0.65  0.97  134      up
 7    hdd  0.09769   1.00000  100 GiB  662 MiB  608 MiB   13 KiB   54 MiB   99 GiB  0.65  0.97  134      up
 8    hdd  0.09769   1.00000  100 GiB  713 MiB  659 MiB   13 KiB   54 MiB   99 GiB  0.70  1.05  133      up
                       TOTAL  900 GiB  6.0 GiB  5.4 GiB  104 KiB  588 MiB  894 GiB  0.67                   
MIN/MAX VAR: 0.67/1.22  STDDEV: 0.08
[root@ceph-deploy ~]#

三、crush 运行图管理

通过工具将 ceph 的 crush 运行图导出并进行编辑, 然后导入
导出 crush 运行图:
注: 导出的 crush 运行图为二进制格式, 无法通过文本编辑器直接打开, 需要使用 crushtool工具转换为文本格式后才能通过 vim 等文本编辑宫工具打开和编辑。

[root@ceph-deploy ~]# mkdir /data/ceph -p
[root@ceph-deploy ~]# ceph osd getcrushmap -o /data/ceph/crushmap-v1
32
[root@ceph-deploy ~]# crushtool -d /data/ceph/crushmap-v1 > /data/ceph/crushmap-v1.txt  #二进制转文本文件
[root@ceph-deploy ~]# file /data/ceph/crushmap-v1.txt 
/data/ceph/crushmap-v1.txt: ASCII text
[root@ceph-deploy ~]# 

[root@ceph-deploy ~]# cat /data/ceph/crushmap-v1.txt 

# begin crush map  #可调整的 crush map 参数
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable chooseleaf_stable 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54

# devices #当前的设备列表
device 0 osd.0 class hdd
device 1 osd.1 class hdd
device 2 osd.2 class hdd
device 3 osd.3 class hdd
device 4 osd.4 class hdd
device 5 osd.5 class hdd
device 6 osd.6 class hdd
device 7 osd.7 class hdd
device 8 osd.8 class hdd

# types #当前支持的 bucket 类型
type 0 osd #osd 守护进程, 对应到一个磁盘设备
type 1 host #一个主机
type 2 chassis #刀片服务器的机箱
type 3 rack #包含若干个服务器的机柜/机架
type 4 row #包含若干个机柜的一排机柜(一行机柜)
type 5 pdu #机柜的接入电源插座
type 6 pod #一个机房中的若干个小房间
type 7 room #包含若干机柜的房间, 一个数据中心有好多这样的房间组成
type 8 datacenter #一个数据中心或 IDC
type 9 zone #区域, 地区
type 10 region #可用域, 比如 AWS 宁夏中卫数据中心
type 11 root #bucket 分层的最顶部, 根

# buckets
host ceph-node1 {  #类型 Host 名称为 ceph-node1
    id -3        # do not change unnecessarily  #ceph 生成的 OSD ID, 非必要不要改
    id -4 class hdd        # do not change unnecessarily
    # weight 0.293  #crush 算法, 管理 OSD 角色
    alg straw2
    hash 0    # rjenkins1  # rjenkins1 #使用是哪个 hash 算法, 0 表示选择 rjenkins1 这种 hash算法
    item osd.0 weight 0.098 #osd0 权重比例, crush 会自动根据磁盘空间计算, 不同的磁盘空间的权重不一样
    item osd.1 weight 0.098
    item osd.2 weight 0.098
}
host ceph-node2 {
    id -5        # do not change unnecessarily
    id -6 class hdd        # do not change unnecessarily
    # weight 0.293
    alg straw2
    hash 0    # rjenkins1
    item osd.3 weight 0.098
    item osd.4 weight 0.098
    item osd.5 weight 0.098
}
host ceph-node3 {
    id -7        # do not change unnecessarily
    id -8 class hdd        # do not change unnecessarily
    # weight 0.293
    alg straw2
    hash 0    # rjenkins1
    item osd.6 weight 0.098
    item osd.7 weight 0.098
    item osd.8 weight 0.098
}
root default { #根的配置
    id -1        # do not change unnecessarily
    id -2 class hdd        # do not change unnecessarily
    # weight 0.879
    alg straw2
    hash 0    # rjenkins1
    item ceph-node1 weight 0.293
    item ceph-node2 weight 0.293
    item ceph-node3 weight 0.293
}

# rules
rule replicated_rule { #副本池的默认配置
    id 0
    type replicated
    min_size 1
    max_size 10 #默认最大副本为 10
    step take default #基于 default 定义的主机分配 OSD
    step chooseleaf firstn 0 type host #选择主机, 故障域类型为主机
    step emit #弹出配置即返回给客户端
}
rule erasure-code { #纠删码池的默认配置
    id 1
    type erasure
    min_size 3
    max_size 4
    step set_chooseleaf_tries 5
    step set_choose_tries 100
    step take default
    step chooseleaf indep 0 type host
    step emit
}

# end crush map

将文本转换为 crush 格式

root@ceph-deploy:~# crushtool -c /data/ceph/crushmap-v1.txt -o /data/ceph/crushmap-v2

导入新的 crush
导入的运行图会立即覆盖原有的运行图并立即生效。

root@ceph-deploy:~# ceph osd setcrushmap -i /data/ceph/crushmap-v2

验证 crush 运行图是否生效:

[root@ceph-deploy ~]# ceph osd crush rule dump

[
    {
        "rule_id": 0,
        "rule_name": "replicated_rule",
        "ruleset": 0,
        "type": 1,
        "min_size": 1,
        "max_size": 8,
        "steps": [
            {
                "op": "take",
                "item": -1,
                "item_name": "default"
            },
            {
                "op": "chooseleaf_firstn",
                "num": 0,
                "type": "host"
            },
            {
                "op": "emit"
            }
        ]
    },
    {
        "rule_id": 1,
        "rule_name": "erasure-code",
        "ruleset": 1,
        "type": 3,
        "min_size": 3,
        "max_size": 4,
        "steps": [
            {
                "op": "set_chooseleaf_tries",
                "num": 5
            },
            {
                "op": "set_choose_tries",
                "num": 100
            },
            {
                "op": "take",
                "item": -1,
                "item_name": "default"
            },
            {
                "op": "chooseleaf_indep",
                "num": 0,
                "type": "host"
            },
            {
                "op": "emit"
            }
        ]
    }
]

四、crush 数据分类管理

Ceph crush 算法分配的 PG 的时候可以将 PG 分配到不同主机的 OSD 上, 以实现以主机为单位的高可用, 这也是默认机制, 但是无法保证不同 PG 位于不同机柜或者机房的主机, 如果要实现基于机柜或者是更高级的 IDC 等方式的数据高可用, 而且也不能实现 A 项目的数据在 SSD, B 项目的数据在机械盘,如果想要实现此功能则需要导出 crush 运行图并手动编辑, 之后再导入并覆盖原有的 crush 运行图。

 

五、集群中ssh、hdd分类管理

模拟集群内,物理机有ssd、hdd两种不同类型磁盘,然后进行分类管理

[root@ceph-deploy ~]# ceph osd getcrushmap -o /data/ceph/crushmap-v3
[root@ceph-deploy ~]# crushtool -d /data/ceph/crushmap-v3 > /data/ceph/crushmap-v3.txt

[root@ceph-deploy ~]# vi /data/ceph/crushmap-v3.txt 

# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable chooseleaf_stable 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54

# devices
device 0 osd.0 class hdd
device 1 osd.1 class hdd
device 2 osd.2 class hdd
device 3 osd.3 class hdd
device 4 osd.4 class hdd
device 5 osd.5 class hdd
device 6 osd.6 class hdd
device 7 osd.7 class hdd
device 8 osd.8 class hdd

# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 zone
type 10 region
type 11 root

# buckets
host ceph-node1 {
    id -3        # do not change unnecessarily
    id -4 class hdd        # do not change unnecessarily
    # weight 0.293
    alg straw2
    hash 0    # rjenkins1
    item osd.0 weight 0.098
    item osd.1 weight 0.098
}
host ceph-node2 {
    id -5        # do not change unnecessarily
    id -6 class hdd        # do not change unnecessarily
    # weight 0.293
    alg straw2
    hash 0    # rjenkins1
    item osd.3 weight 0.098
    item osd.4 weight 0.098
}
host ceph-node3 {
    id -7        # do not change unnecessarily
    id -8 class hdd        # do not change unnecessarily
    # weight 0.293
    alg straw2
    hash 0    # rjenkins1
    item osd.6 weight 0.098
    item osd.7 weight 0.098
}
root default {
    id -1        # do not change unnecessarily
    id -2 class hdd        # do not change unnecessarily
    # weight 0.879
    alg straw2
    hash 0    # rjenkins1
    item ceph-node1 weight 0.293
    item ceph-node2 weight 0.293
    item ceph-node3 weight 0.293
}

#ssd node
host ceph-ssd-node1 {
        id -13           # do not change unnecessarily
        id -14 class ssd         # do not change unnecessarily
        # weight 0.293
        alg straw2
        hash 0  # rjenkins1
        item osd.2 weight 0.098
}

host ceph-ssd-node2 {
        id -15           # do not change unnecessarily
        id -16 class ssd         # do not change unnecessarily
        # weight 0.293
        alg straw2
        hash 0  # rjenkins1
        item osd.5 weight 0.098
}
host ceph-ssd-node3 {
        id -17           # do not change unnecessarily
        id -18 class ssd         # do not change unnecessarily
        # weight 0.293
        alg straw2
        hash 0  # rjenkins1
        item osd.8 weight 0.098
}

root ssd {
        id -11           # do not change unnecessarily
        id -12 class hdd         # do not change unnecessarily   #真实环境这里需要写真是ssd类型
        # weight 0.879
        alg straw2
        hash 0  # rjenkins1
        item ceph-ssd-node1 weight 0.293
        item ceph-ssd-node2 weight 0.293
        item ceph-ssd-node3 weight 0.293
}


# rules
rule replicated_rule {
    id 0
    type replicated
    min_size 1
    max_size 10
    step take default
    step chooseleaf firstn 0 type host
    step emit
}

rule ssd_rule {
        id 10
        type replicated
        min_size 1
        max_size 5
        step take ssd
        step chooseleaf firstn 0 type host
        step emit
}

rule erasure-code {
    id 1
    type erasure
    min_size 3
    max_size 4
    step set_chooseleaf_tries 5
    step set_choose_tries 100
    step take default
    step chooseleaf indep 0 type host
    step emit
}

# end crush map
[root@ceph-deploy ~]# 

将v3文本文件编译成v4二进制文件导入

[root@ceph-deploy ~]# crushtool -c /data/ceph/crushmap-v3.txt -o /data/ceph/crushmap-v4
[root@ceph-deploy ~]# ceph osd setcrushmap -i /data/ceph/crushmap-v4
33
[root@ceph-deploy ~]# 

[root@ceph-deploy ~]# ceph osd crush rule dump

[
    {
        "rule_id": 0,
        "rule_name": "replicated_rule",
        "ruleset": 0,
        "type": 1,
        "min_size": 1,
        "max_size": 10,
        "steps": [
            {
                "op": "take",
                "item": -1,
                "item_name": "default"
            },
            {
                "op": "chooseleaf_firstn",
                "num": 0,
                "type": "host"
            },
            {
                "op": "emit"
            }
        ]
    },
    {
        "rule_id": 1,
        "rule_name": "erasure-code",
        "ruleset": 1,
        "type": 3,
        "min_size": 3,
        "max_size": 4,
        "steps": [
            {
                "op": "set_chooseleaf_tries",
                "num": 5
            },
            {
                "op": "set_choose_tries",
                "num": 100
            },
            {
                "op": "take",
                "item": -1,
                "item_name": "default"
            },
            {
                "op": "chooseleaf_indep",
                "num": 0,
                "type": "host"
            },
            {
                "op": "emit"
            }
        ]
    },
    {
        "rule_id": 10,
        "rule_name": "ssd_rule",
        "ruleset": 10,
        "type": 1,
        "min_size": 1,
        "max_size": 5,
        "steps": [
            {
                "op": "take",
                "item": -11,
                "item_name": "ssd"
            },
            {
                "op": "chooseleaf_firstn",
                "num": 0,
                "type": "host"
            },
            {
                "op": "emit"
            }
        ]
    }
]

查看集群状态

[root@ceph-deploy ~]# ceph osd df
ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS
 2    hdd  0.09799   1.00000  100 GiB   63 MiB   10 MiB    6 KiB   53 MiB  100 GiB  0.06  0.09    0      up #分类后的osd pgs会自动移除
 5    hdd  0.09799   1.00000  100 GiB   87 MiB   10 MiB    8 KiB   77 MiB  100 GiB  0.09  0.13    0      up
 8    hdd  0.09799   1.00000  100 GiB   78 MiB   10 MiB   13 KiB   68 MiB  100 GiB  0.08  0.11    0      up
 0    hdd  0.09799   1.00000  100 GiB  989 MiB  911 MiB   20 KiB   78 MiB   99 GiB  0.97  1.42  197      up
 1    hdd  0.09799   1.00000  100 GiB  1.0 GiB  932 MiB    7 KiB  111 MiB   99 GiB  1.02  1.50  204      up
 3    hdd  0.09799   1.00000  100 GiB  881 MiB  816 MiB   12 KiB   65 MiB   99 GiB  0.86  1.27  200      up
 4    hdd  0.09799   1.00000  100 GiB  1.1 GiB  1.0 GiB   13 KiB   78 MiB   99 GiB  1.08  1.59  201      up
 6    hdd  0.09799   1.00000  100 GiB  1.0 GiB  967 MiB    7 KiB   91 MiB   99 GiB  1.03  1.52  197      up
 7    hdd  0.09799   1.00000  100 GiB  949 MiB  876 MiB   13 KiB   73 MiB   99 GiB  0.93  1.37  204      up
                       TOTAL  900 GiB  6.1 GiB  5.4 GiB  104 KiB  694 MiB  894 GiB  0.68                   
MIN/MAX VAR: 0.09/1.59  STDDEV: 0.43
[root@ceph-deploy ~]# ceph osd df tree
ID   CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              
-11         0.87900         -  300 GiB  228 MiB   30 MiB   27 KiB  198 MiB  300 GiB  0.07  0.11    -          root ssd               
-13         0.29300         -  100 GiB   63 MiB   10 MiB    6 KiB   53 MiB  100 GiB  0.06  0.09    -              host ceph-ssd-node1
  2    hdd  0.09799   1.00000  100 GiB   63 MiB   10 MiB    6 KiB   53 MiB  100 GiB  0.06  0.09    0      up          osd.2          
-15         0.29300         -  100 GiB   87 MiB   10 MiB    8 KiB   77 MiB  100 GiB  0.09  0.13    -              host ceph-ssd-node2
  5    hdd  0.09799   1.00000  100 GiB   87 MiB   10 MiB    8 KiB   77 MiB  100 GiB  0.09  0.13    0      up          osd.5          
-17         0.29300         -  100 GiB   78 MiB   10 MiB   13 KiB   68 MiB  100 GiB  0.08  0.11    -              host ceph-ssd-node3
  8    hdd  0.09799   1.00000  100 GiB   78 MiB   10 MiB   13 KiB   68 MiB  100 GiB  0.08  0.11    0      up          osd.8          
 -1         0.87900         -  600 GiB  5.9 GiB  5.4 GiB   72 KiB  500 MiB  594 GiB  0.98  1.45    -          root default           
 -3         0.29300         -  200 GiB  2.0 GiB  1.8 GiB   27 KiB  189 MiB  198 GiB  0.99  1.46    -              host ceph-node1    
  0    hdd  0.09799   1.00000  100 GiB  990 MiB  911 MiB   20 KiB   78 MiB   99 GiB  0.97  1.42  197      up          osd.0          
  1    hdd  0.09799   1.00000  100 GiB  1.0 GiB  932 MiB    7 KiB  111 MiB   99 GiB  1.02  1.50  204      up          osd.1          
 -5         0.29300         -  200 GiB  1.9 GiB  1.8 GiB   25 KiB  147 MiB  198 GiB  0.97  1.43    -              host ceph-node2    
  3    hdd  0.09799   1.00000  100 GiB  885 MiB  816 MiB   12 KiB   69 MiB   99 GiB  0.86  1.27  200      up          osd.3          
  4    hdd  0.09799   1.00000  100 GiB  1.1 GiB  1.0 GiB   13 KiB   78 MiB   99 GiB  1.08  1.59  201      up          osd.4          
 -7         0.29300         -  200 GiB  2.0 GiB  1.8 GiB   20 KiB  163 MiB  198 GiB  0.98  1.44    -              host ceph-node3    
  6    hdd  0.09799   1.00000  100 GiB  1.0 GiB  967 MiB    7 KiB   91 MiB   99 GiB  1.03  1.52  197      up          osd.6          
  7    hdd  0.09799   1.00000  100 GiB  949 MiB  876 MiB   13 KiB   73 MiB   99 GiB  0.93  1.36  204      up          osd.7          
                        TOTAL  900 GiB  6.1 GiB  5.4 GiB  104 KiB  698 MiB  894 GiB  0.68                                            
MIN/MAX VAR: 0.09/1.59  STDDEV: 0.43
[root@ceph-deploy ~]# ceph -s
  cluster:
    id:     8dc32c41-121c-49df-9554-dfb7deb8c975
    health: HEALTH_OK
 
  services:
    mon: 3 daemons, quorum ceph-mon1,ceph-mon2,ceph-mon3 (age 23h)
    mgr: ceph-mgr1(active, since 23h), standbys: ceph-mgr2
    mds: 2/2 daemons up, 2 standby
    osd: 9 osds: 9 up (since 23h), 9 in (since 7d)
    rgw: 2 daemons active (2 hosts, 1 zones)
 
  data:
    volumes: 1/1 healthy
    pools:   12 pools, 401 pgs
    objects: 843 objects, 1.8 GiB
    usage:   6.1 GiB used, 894 GiB / 900 GiB avail
    pgs:     401 active+clean
 
[root@ceph-deploy ~]#

再次导出验证crush map文件(一般主体文件不写错,次要配置ceph会自动帮忙修正)

[root@ceph-deploy ~]# ceph osd getcrushmap -o /data/ceph/crushmap-v5
33
[root@ceph-deploy ~]# crushtool -d /data/ceph/crushmap-v5 > /data/ceph/crushmap-v5.txt
[root@ceph-deploy ~]# 

[root@ceph-deploy ceph]# cat crushmap-v5.txt

# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable chooseleaf_stable 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54

# devices
device 0 osd.0 class hdd
device 1 osd.1 class hdd
device 2 osd.2 class hdd
device 3 osd.3 class hdd
device 4 osd.4 class hdd
device 5 osd.5 class hdd
device 6 osd.6 class hdd
device 7 osd.7 class hdd
device 8 osd.8 class hdd

# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 zone
type 10 region
type 11 root

# buckets
host ceph-node1 {
    id -3        # do not change unnecessarily
    id -4 class hdd        # do not change unnecessarily
    id -21 class ssd        # do not change unnecessarily
    # weight 0.196
    alg straw2
    hash 0    # rjenkins1
    item osd.0 weight 0.098
    item osd.1 weight 0.098
}
host ceph-node2 {
    id -5        # do not change unnecessarily
    id -6 class hdd        # do not change unnecessarily
    id -22 class ssd        # do not change unnecessarily
    # weight 0.196
    alg straw2
    hash 0    # rjenkins1
    item osd.3 weight 0.098
    item osd.4 weight 0.098
}
host ceph-node3 {
    id -7        # do not change unnecessarily
    id -8 class hdd        # do not change unnecessarily
    id -23 class ssd        # do not change unnecessarily
    # weight 0.196
    alg straw2
    hash 0    # rjenkins1
    item osd.6 weight 0.098
    item osd.7 weight 0.098
}
root default {
    id -1        # do not change unnecessarily
    id -2 class hdd        # do not change unnecessarily
    id -24 class ssd        # do not change unnecessarily
    # weight 0.879
    alg straw2
    hash 0    # rjenkins1
    item ceph-node1 weight 0.293
    item ceph-node2 weight 0.293
    item ceph-node3 weight 0.293
}
host ceph-ssd-node1 {
    id -13        # do not change unnecessarily
    id -9 class hdd        # do not change unnecessarily
    id -14 class ssd        # do not change unnecessarily
    # weight 0.098
    alg straw2
    hash 0    # rjenkins1
    item osd.2 weight 0.098
}
host ceph-ssd-node2 {
    id -15        # do not change unnecessarily
    id -10 class hdd        # do not change unnecessarily
    id -16 class ssd        # do not change unnecessarily
    # weight 0.098
    alg straw2
    hash 0    # rjenkins1
    item osd.5 weight 0.098
}
host ceph-ssd-node3 {
    id -17        # do not change unnecessarily
    id -19 class hdd        # do not change unnecessarily
    id -18 class ssd        # do not change unnecessarily
    # weight 0.098
    alg straw2
    hash 0    # rjenkins1
    item osd.8 weight 0.098
}
root ssd {
    id -11        # do not change unnecessarily
    id -12 class hdd        # do not change unnecessarily
    id -20 class ssd        # do not change unnecessarily
    # weight 0.879
    alg straw2
    hash 0    # rjenkins1
    item ceph-ssd-node1 weight 0.293
    item ceph-ssd-node2 weight 0.293
    item ceph-ssd-node3 weight 0.293
}

# rules
rule replicated_rule {
    id 0
    type replicated
    min_size 1
    max_size 10
    step take default
    step chooseleaf firstn 0 type host
    step emit
}
rule erasure-code {
    id 1
    type erasure
    min_size 3
    max_size 4
    step set_chooseleaf_tries 5
    step set_choose_tries 100
    step take default
    step chooseleaf indep 0 type host
    step emit
}
rule ssd_rule {
    id 10
    type replicated
    min_size 1
    max_size 5
    step take ssd
    step chooseleaf firstn 0 type host
    step emit
}

# end crush map
View Code

测试创建存储池:

[root@ceph-deploy ~]# ceph osd pool create ssd-pool 32 32 ssd_rule
pool 'ssd-pool' created
[root@ceph-deploy ~]#
[root@ceph-deploy ~]# ceph osd  pool ls
device_health_metrics
myrbd1
.rgw.root
default.rgw.log
default.rgw.control
default.rgw.meta
cephfs-metadata
cephfs-data
mypool
rbd-data1
default.rgw.buckets.index
default.rgw.buckets.data
ssd-pool
[root@ceph-deploy ~]# 

验证 pgp 状态:

[root@ceph-deploy ~]# ceph pg ls-by-pool ssd-pool | awk '{print $1,$2,$15}'  
PG OBJECTS ACTING
16.0 0 [8,5,2]p8
16.1 0 [2,5,8]p2
16.2 0 [2,8,5]p2
16.3 0 [2,5,8]p2
16.4 0 [5,8,2]p5
16.5 0 [8,2,5]p8
16.6 0 [2,8,5]p2
16.7 0 [8,2,5]p8
16.8 0 [5,2,8]p5
16.9 0 [8,2,5]p8
16.a 0 [2,8,5]p2
16.b 0 [8,2,5]p8
16.c 0 [8,5,2]p8
16.d 0 [5,2,8]p5
16.e 0 [8,2,5]p8
16.f 0 [8,2,5]p8
16.10 0 [8,5,2]p8
16.11 0 [8,5,2]p8
16.12 0 [8,5,2]p8
16.13 0 [8,5,2]p8
16.14 0 [2,5,8]p2
16.15 0 [8,5,2]p8
16.16 0 [5,2,8]p5
16.17 0 [2,8,5]p2
16.18 0 [2,5,8]p2
16.19 0 [5,2,8]p5
16.1a 0 [2,8,5]p2
16.1b 0 [5,8,2]p5
16.1c 0 [2,8,5]p2
16.1d 0 [5,8,2]p5
16.1e 0 [2,5,8]p2
16.1f 0 [8,5,2]p8
  
* NOTE: afterwards
[root@ceph-deploy ~]# 

真是生产环境中ssd、hdd节点分类管理情形

# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable chooseleaf_stable 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54

# devices
device 0 osd.0 class hdd
device 1 osd.1 class hdd
device 2 osd.2 class hdd
device 3 osd.3 class hdd
device 4 osd.4 class hdd
device 5 osd.5 class hdd
device 6 osd.6 class hdd
device 7 osd.7 class hdd
device 8 osd.8 class hdd
device 9 osd.9 class hdd
device 10 osd.10 class hdd
device 11 osd.11 class hdd
device 12 osd.12 class hdd
device 13 osd.13 class hdd
device 14 osd.14 class hdd
device 20 osd.20 class hdd
device 21 osd.21 class hdd
device 22 osd.22 class hdd
device 23 osd.23 class hdd
device 24 osd.24 class hdd
device 25 osd.25 class hdd
device 26 osd.26 class hdd
device 27 osd.27 class hdd
device 28 osd.28 class hdd
device 29 osd.29 class hdd
device 30 osd.30 class hdd
device 31 osd.31 class hdd
device 32 osd.32 class hdd
device 33 osd.33 class hdd
device 34 osd.34 class hdd
device 40 osd.40 class hdd
device 41 osd.41 class hdd
device 42 osd.42 class hdd
device 43 osd.43 class hdd
device 44 osd.44 class hdd
device 45 osd.45 class hdd
device 46 osd.46 class hdd
device 47 osd.47 class hdd
device 48 osd.48 class hdd
device 49 osd.49 class hdd
device 50 osd.50 class hdd
device 51 osd.51 class hdd
device 52 osd.52 class hdd
device 53 osd.53 class hdd
device 54 osd.54 class hdd
device 60 osd.60 class hdd
device 61 osd.61 class hdd
device 62 osd.62 class hdd
device 63 osd.63 class hdd
device 64 osd.64 class hdd
device 65 osd.65 class hdd
device 66 osd.66 class hdd
device 67 osd.67 class hdd
device 68 osd.68 class hdd
device 69 osd.69 class hdd
device 70 osd.70 class hdd
device 71 osd.71 class hdd
device 72 osd.72 class hdd
device 73 osd.73 class hdd
device 74 osd.74 class hdd
device 80 osd.80 class hdd
device 81 osd.81 class hdd
device 82 osd.82 class hdd
device 83 osd.83 class hdd
device 84 osd.84 class hdd
device 85 osd.85 class hdd
device 86 osd.86 class hdd
device 87 osd.87 class hdd
device 88 osd.88 class hdd
device 89 osd.89 class hdd
device 90 osd.90 class hdd
device 91 osd.91 class hdd
device 92 osd.92 class hdd
device 93 osd.93 class hdd
device 94 osd.94 class hdd
device 100 osd.100 class hdd
device 101 osd.101 class hdd
device 102 osd.102 class hdd
device 103 osd.103 class hdd
device 104 osd.104 class hdd
device 105 osd.105 class hdd
device 106 osd.106 class hdd
device 107 osd.107 class hdd
device 108 osd.108 class hdd
device 109 osd.109 class hdd
device 110 osd.110 class hdd
device 111 osd.111 class hdd
device 112 osd.112 class hdd
device 113 osd.113 class hdd
device 114 osd.114 class hdd
device 120 osd.120 class hdd
device 121 osd.121 class hdd
device 122 osd.122 class hdd
device 123 osd.123 class hdd
device 124 osd.124 class hdd
device 125 osd.125 class hdd
device 126 osd.126 class hdd
device 127 osd.127 class hdd
device 128 osd.128 class hdd
device 129 osd.129 class hdd
device 130 osd.130 class hdd
device 131 osd.131 class hdd
device 132 osd.132 class hdd
device 133 osd.133 class hdd
device 134 osd.134 class hdd
device 140 osd.140 class hdd
device 141 osd.141 class hdd
device 142 osd.142 class hdd
device 143 osd.143 class hdd
device 144 osd.144 class hdd
device 145 osd.145 class hdd
device 146 osd.146 class hdd
device 147 osd.147 class hdd
device 148 osd.148 class hdd
device 149 osd.149 class hdd
device 150 osd.150 class hdd
device 151 osd.151 class hdd
device 152 osd.152 class hdd
device 153 osd.153 class hdd
device 154 osd.154 class hdd
device 160 osd.160 class hdd
device 161 osd.161 class hdd
device 162 osd.162 class hdd
device 163 osd.163 class hdd
device 164 osd.164 class hdd
device 165 osd.165 class hdd
device 166 osd.166 class hdd
device 167 osd.167 class hdd
device 168 osd.168 class hdd
device 169 osd.169 class hdd
device 170 osd.170 class hdd
device 171 osd.171 class hdd
device 172 osd.172 class hdd
device 173 osd.173 class hdd
device 174 osd.174 class hdd
device 180 osd.180 class hdd
device 181 osd.181 class hdd
device 182 osd.182 class hdd
device 183 osd.183 class hdd
device 184 osd.184 class hdd
device 185 osd.185 class hdd
device 186 osd.186 class hdd
device 187 osd.187 class hdd
device 188 osd.188 class hdd
device 189 osd.189 class hdd
device 190 osd.190 class hdd
device 191 osd.191 class hdd
device 192 osd.192 class hdd
device 193 osd.193 class hdd
device 194 osd.194 class hdd
device 200 osd.200 class hdd
device 201 osd.201 class hdd
device 202 osd.202 class hdd
device 203 osd.203 class hdd
device 204 osd.204 class hdd
device 205 osd.205 class hdd
device 206 osd.206 class hdd
device 207 osd.207 class hdd
device 208 osd.208 class hdd
device 209 osd.209 class hdd
device 210 osd.210 class hdd
device 211 osd.211 class hdd
device 212 osd.212 class hdd
device 213 osd.213 class hdd
device 214 osd.214 class hdd
device 220 osd.220 class hdd
device 221 osd.221 class hdd
device 222 osd.222 class hdd
device 223 osd.223 class hdd
device 224 osd.224 class hdd
device 225 osd.225 class hdd
device 226 osd.226 class hdd
device 227 osd.227 class hdd
device 228 osd.228 class hdd
device 229 osd.229 class hdd
device 230 osd.230 class hdd
device 231 osd.231 class hdd
device 232 osd.232 class hdd
device 233 osd.233 class hdd
device 234 osd.234 class hdd
device 240 osd.240 class hdd
device 241 osd.241 class hdd
device 242 osd.242 class hdd
device 243 osd.243 class hdd
device 244 osd.244 class hdd
device 245 osd.245 class hdd
device 246 osd.246 class hdd
device 247 osd.247 class hdd
device 248 osd.248 class hdd
device 249 osd.249 class hdd
device 250 osd.250 class hdd
device 251 osd.251 class hdd
device 252 osd.252 class hdd
device 253 osd.253 class hdd
device 254 osd.254 class hdd
device 260 osd.260 class hdd
device 261 osd.261 class hdd
device 262 osd.262 class hdd
device 263 osd.263 class hdd
device 264 osd.264 class hdd
device 265 osd.265 class hdd
device 266 osd.266 class hdd
device 267 osd.267 class hdd
device 268 osd.268 class hdd
device 269 osd.269 class hdd
device 270 osd.270 class hdd
device 271 osd.271 class hdd
device 272 osd.272 class hdd
device 273 osd.273 class hdd
device 274 osd.274 class hdd
device 280 osd.280 class hdd
device 281 osd.281 class hdd
device 282 osd.282 class hdd
device 283 osd.283 class hdd
device 284 osd.284 class hdd
device 285 osd.285 class hdd
device 286 osd.286 class hdd
device 287 osd.287 class hdd
device 288 osd.288 class hdd
device 289 osd.289 class hdd
device 290 osd.290 class hdd
device 291 osd.291 class hdd
device 292 osd.292 class hdd
device 293 osd.293 class hdd
device 294 osd.294 class hdd
device 300 osd.300 class hdd
device 301 osd.301 class hdd
device 302 osd.302 class hdd
device 303 osd.303 class hdd
device 304 osd.304 class hdd
device 305 osd.305 class hdd
device 306 osd.306 class hdd
device 307 osd.307 class hdd
device 308 osd.308 class hdd
device 309 osd.309 class hdd
device 310 osd.310 class hdd
device 311 osd.311 class hdd
device 312 osd.312 class hdd
device 313 osd.313 class hdd
device 314 osd.314 class hdd
device 320 osd.320 class ssd
device 321 osd.321 class ssd
device 322 osd.322 class ssd
device 323 osd.323 class ssd
device 324 osd.324 class ssd
device 325 osd.325 class ssd
device 326 osd.326 class ssd
device 327 osd.327 class ssd
device 328 osd.328 class ssd
device 329 osd.329 class ssd
device 330 osd.330 class ssd
device 331 osd.331 class ssd
device 332 osd.332 class ssd
device 333 osd.333 class ssd
device 334 osd.334 class ssd
device 335 osd.335 class ssd
device 336 osd.336 class ssd
device 337 osd.337 class ssd
device 338 osd.338 class ssd
device 339 osd.339 class ssd
device 340 osd.340 class ssd
device 341 osd.341 class ssd
device 342 osd.342 class ssd
device 343 osd.343 class ssd
device 344 osd.344 class ssd
device 345 osd.345 class ssd
device 346 osd.346 class ssd
device 347 osd.347 class ssd
device 348 osd.348 class ssd
device 349 osd.349 class ssd
device 350 osd.350 class ssd
device 351 osd.351 class ssd
device 352 osd.352 class ssd
device 353 osd.353 class ssd
device 354 osd.354 class ssd
device 355 osd.355 class ssd
device 356 osd.356 class ssd
device 357 osd.357 class ssd
device 358 osd.358 class ssd
device 359 osd.359 class ssd
device 360 osd.360 class ssd
device 361 osd.361 class ssd
device 362 osd.362 class ssd
device 363 osd.363 class ssd
device 364 osd.364 class ssd
device 365 osd.365 class ssd
device 366 osd.366 class ssd
device 367 osd.367 class ssd
device 368 osd.368 class ssd
device 369 osd.369 class ssd
device 370 osd.370 class ssd
device 371 osd.371 class ssd
device 372 osd.372 class ssd
device 373 osd.373 class ssd
device 374 osd.374 class ssd
device 375 osd.375 class ssd
device 376 osd.376 class ssd
device 377 osd.377 class ssd
device 378 osd.378 class ssd
device 379 osd.379 class ssd
device 380 osd.380 class ssd
device 381 osd.381 class ssd
device 382 osd.382 class ssd
device 383 osd.383 class ssd
device 384 osd.384 class ssd
device 385 osd.385 class ssd
device 386 osd.386 class ssd
device 387 osd.387 class ssd
device 388 osd.388 class ssd
device 389 osd.389 class ssd
device 390 osd.390 class ssd
device 391 osd.391 class ssd
device 392 osd.392 class ssd
device 393 osd.393 class ssd
device 394 osd.394 class ssd
device 395 osd.395 class ssd
device 396 osd.396 class ssd
device 397 osd.397 class ssd
device 398 osd.398 class ssd
device 399 osd.399 class ssd
device 400 osd.400 class ssd
device 401 osd.401 class ssd
device 402 osd.402 class ssd
device 403 osd.403 class ssd
device 404 osd.404 class ssd
device 405 osd.405 class ssd
device 406 osd.406 class ssd
device 407 osd.407 class ssd
device 408 osd.408 class ssd
device 409 osd.409 class ssd
device 410 osd.410 class ssd
device 411 osd.411 class ssd
device 412 osd.412 class ssd
device 413 osd.413 class ssd
device 414 osd.414 class ssd
device 415 osd.415 class ssd
device 416 osd.416 class ssd
device 417 osd.417 class ssd
device 418 osd.418 class ssd
device 419 osd.419 class ssd
device 420 osd.420 class ssd
device 421 osd.421 class ssd
device 422 osd.422 class ssd
device 423 osd.423 class ssd
device 424 osd.424 class ssd
device 425 osd.425 class ssd
device 426 osd.426 class ssd
device 427 osd.427 class ssd
device 428 osd.428 class ssd
device 429 osd.429 class ssd
device 430 osd.430 class ssd
device 431 osd.431 class ssd
device 432 osd.432 class ssd
device 433 osd.433 class ssd
device 434 osd.434 class ssd
device 435 osd.435 class ssd
device 436 osd.436 class ssd
device 437 osd.437 class ssd
device 438 osd.438 class ssd
device 439 osd.439 class ssd

# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 region
type 10 root

# buckets
host dt-1ap213-proxmox-01 {
    id -3        # do not change unnecessarily
    id -4 class hdd        # do not change unnecessarily
    id -35 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.0 weight 2.183
    item osd.1 weight 2.183
    item osd.2 weight 2.183
    item osd.3 weight 2.183
    item osd.4 weight 2.183
    item osd.5 weight 2.183
    item osd.6 weight 2.183
    item osd.7 weight 2.183
    item osd.8 weight 2.183
    item osd.9 weight 2.183
    item osd.10 weight 2.183
    item osd.11 weight 2.183
    item osd.12 weight 2.183
    item osd.13 weight 2.183
    item osd.14 weight 2.183
}
host dt-1ap213-proxmox-02 {
    id -5        # do not change unnecessarily
    id -6 class hdd        # do not change unnecessarily
    id -36 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.20 weight 2.183
    item osd.21 weight 2.183
    item osd.22 weight 2.183
    item osd.23 weight 2.183
    item osd.24 weight 2.183
    item osd.25 weight 2.183
    item osd.26 weight 2.183
    item osd.27 weight 2.183
    item osd.28 weight 2.183
    item osd.29 weight 2.183
    item osd.30 weight 2.183
    item osd.31 weight 2.183
    item osd.32 weight 2.183
    item osd.33 weight 2.183
    item osd.34 weight 2.183
}
host dt-1ap213-proxmox-03 {
    id -7        # do not change unnecessarily
    id -8 class hdd        # do not change unnecessarily
    id -37 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.40 weight 2.183
    item osd.41 weight 2.183
    item osd.42 weight 2.183
    item osd.43 weight 2.183
    item osd.44 weight 2.183
    item osd.45 weight 2.183
    item osd.46 weight 2.183
    item osd.47 weight 2.183
    item osd.48 weight 2.183
    item osd.49 weight 2.183
    item osd.50 weight 2.183
    item osd.51 weight 2.183
    item osd.52 weight 2.183
    item osd.53 weight 2.183
    item osd.54 weight 2.183
}
host dt-1ap213-proxmox-04 {
    id -9        # do not change unnecessarily
    id -10 class hdd        # do not change unnecessarily
    id -38 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.60 weight 2.183
    item osd.61 weight 2.183
    item osd.62 weight 2.183
    item osd.63 weight 2.183
    item osd.64 weight 2.183
    item osd.65 weight 2.183
    item osd.66 weight 2.183
    item osd.67 weight 2.183
    item osd.68 weight 2.183
    item osd.69 weight 2.183
    item osd.70 weight 2.183
    item osd.71 weight 2.183
    item osd.72 weight 2.183
    item osd.73 weight 2.183
    item osd.74 weight 2.183
}
host dt-1ap213-proxmox-05 {
    id -11        # do not change unnecessarily
    id -12 class hdd        # do not change unnecessarily
    id -39 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.80 weight 2.183
    item osd.81 weight 2.183
    item osd.82 weight 2.183
    item osd.83 weight 2.183
    item osd.84 weight 2.183
    item osd.85 weight 2.183
    item osd.86 weight 2.183
    item osd.87 weight 2.183
    item osd.88 weight 2.183
    item osd.89 weight 2.183
    item osd.90 weight 2.183
    item osd.91 weight 2.183
    item osd.92 weight 2.183
    item osd.93 weight 2.183
    item osd.94 weight 2.183
}
host dt-1ap213-proxmox-06 {
    id -13        # do not change unnecessarily
    id -14 class hdd        # do not change unnecessarily
    id -40 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.100 weight 2.183
    item osd.101 weight 2.183
    item osd.102 weight 2.183
    item osd.103 weight 2.183
    item osd.104 weight 2.183
    item osd.105 weight 2.183
    item osd.106 weight 2.183
    item osd.107 weight 2.183
    item osd.108 weight 2.183
    item osd.109 weight 2.183
    item osd.110 weight 2.183
    item osd.111 weight 2.183
    item osd.112 weight 2.183
    item osd.113 weight 2.183
    item osd.114 weight 2.183
}
host dt-1ap213-proxmox-07 {
    id -15        # do not change unnecessarily
    id -16 class hdd        # do not change unnecessarily
    id -41 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.120 weight 2.183
    item osd.121 weight 2.183
    item osd.122 weight 2.183
    item osd.123 weight 2.183
    item osd.124 weight 2.183
    item osd.125 weight 2.183
    item osd.126 weight 2.183
    item osd.127 weight 2.183
    item osd.128 weight 2.183
    item osd.129 weight 2.183
    item osd.130 weight 2.183
    item osd.131 weight 2.183
    item osd.132 weight 2.183
    item osd.133 weight 2.183
    item osd.134 weight 2.183
}
host dt-1ap213-proxmox-08 {
    id -17        # do not change unnecessarily
    id -18 class hdd        # do not change unnecessarily
    id -42 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.140 weight 2.183
    item osd.141 weight 2.183
    item osd.142 weight 2.183
    item osd.143 weight 2.183
    item osd.144 weight 2.183
    item osd.145 weight 2.183
    item osd.146 weight 2.183
    item osd.147 weight 2.183
    item osd.148 weight 2.183
    item osd.149 weight 2.183
    item osd.150 weight 2.183
    item osd.151 weight 2.183
    item osd.152 weight 2.183
    item osd.153 weight 2.183
    item osd.154 weight 2.183
}
host dt-1ap214-proxmox-01 {
    id -19        # do not change unnecessarily
    id -20 class hdd        # do not change unnecessarily
    id -43 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.160 weight 2.183
    item osd.161 weight 2.183
    item osd.162 weight 2.183
    item osd.163 weight 2.183
    item osd.164 weight 2.183
    item osd.165 weight 2.183
    item osd.166 weight 2.183
    item osd.167 weight 2.183
    item osd.168 weight 2.183
    item osd.169 weight 2.183
    item osd.170 weight 2.183
    item osd.171 weight 2.183
    item osd.172 weight 2.183
    item osd.173 weight 2.183
    item osd.174 weight 2.183
}
host dt-1ap214-proxmox-02 {
    id -21        # do not change unnecessarily
    id -22 class hdd        # do not change unnecessarily
    id -44 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.180 weight 2.183
    item osd.181 weight 2.183
    item osd.182 weight 2.183
    item osd.183 weight 2.183
    item osd.184 weight 2.183
    item osd.185 weight 2.183
    item osd.186 weight 2.183
    item osd.187 weight 2.183
    item osd.188 weight 2.183
    item osd.189 weight 2.183
    item osd.190 weight 2.183
    item osd.191 weight 2.183
    item osd.192 weight 2.183
    item osd.193 weight 2.183
    item osd.194 weight 2.183
}
host dt-1ap214-proxmox-03 {
    id -23        # do not change unnecessarily
    id -24 class hdd        # do not change unnecessarily
    id -45 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.200 weight 2.183
    item osd.201 weight 2.183
    item osd.202 weight 2.183
    item osd.203 weight 2.183
    item osd.204 weight 2.183
    item osd.205 weight 2.183
    item osd.206 weight 2.183
    item osd.207 weight 2.183
    item osd.208 weight 2.183
    item osd.209 weight 2.183
    item osd.210 weight 2.183
    item osd.211 weight 2.183
    item osd.212 weight 2.183
    item osd.213 weight 2.183
    item osd.214 weight 2.183
}
host dt-1ap214-proxmox-04 {
    id -25        # do not change unnecessarily
    id -26 class hdd        # do not change unnecessarily
    id -46 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.220 weight 2.183
    item osd.221 weight 2.183
    item osd.222 weight 2.183
    item osd.223 weight 2.183
    item osd.224 weight 2.183
    item osd.225 weight 2.183
    item osd.226 weight 2.183
    item osd.227 weight 2.183
    item osd.228 weight 2.183
    item osd.229 weight 2.183
    item osd.230 weight 2.183
    item osd.231 weight 2.183
    item osd.232 weight 2.183
    item osd.233 weight 2.183
    item osd.234 weight 2.183
}
host dt-1ap214-proxmox-05 {
    id -27        # do not change unnecessarily
    id -28 class hdd        # do not change unnecessarily
    id -47 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.240 weight 2.183
    item osd.241 weight 2.183
    item osd.242 weight 2.183
    item osd.243 weight 2.183
    item osd.244 weight 2.183
    item osd.245 weight 2.183
    item osd.246 weight 2.183
    item osd.247 weight 2.183
    item osd.248 weight 2.183
    item osd.249 weight 2.183
    item osd.250 weight 2.183
    item osd.251 weight 2.183
    item osd.252 weight 2.183
    item osd.253 weight 2.183
    item osd.254 weight 2.183
}
host dt-1ap214-proxmox-06 {
    id -29        # do not change unnecessarily
    id -30 class hdd        # do not change unnecessarily
    id -48 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.260 weight 2.183
    item osd.261 weight 2.183
    item osd.262 weight 2.183
    item osd.263 weight 2.183
    item osd.264 weight 2.183
    item osd.265 weight 2.183
    item osd.266 weight 2.183
    item osd.267 weight 2.183
    item osd.268 weight 2.183
    item osd.269 weight 2.183
    item osd.270 weight 2.183
    item osd.271 weight 2.183
    item osd.272 weight 2.183
    item osd.273 weight 2.183
    item osd.274 weight 2.183
}
host dt-1ap214-proxmox-07 {
    id -31        # do not change unnecessarily
    id -32 class hdd        # do not change unnecessarily
    id -49 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.280 weight 2.183
    item osd.281 weight 2.183
    item osd.282 weight 2.183
    item osd.283 weight 2.183
    item osd.284 weight 2.183
    item osd.285 weight 2.183
    item osd.286 weight 2.183
    item osd.287 weight 2.183
    item osd.288 weight 2.183
    item osd.289 weight 2.183
    item osd.290 weight 2.183
    item osd.291 weight 2.183
    item osd.292 weight 2.183
    item osd.293 weight 2.183
    item osd.294 weight 2.183
}
host dt-1ap214-proxmox-08 {
    id -33        # do not change unnecessarily
    id -34 class hdd        # do not change unnecessarily
    id -50 class ssd        # do not change unnecessarily
    # weight 32.745
    alg straw2
    hash 0    # rjenkins1
    item osd.300 weight 2.183
    item osd.301 weight 2.183
    item osd.302 weight 2.183
    item osd.303 weight 2.183
    item osd.304 weight 2.183
    item osd.305 weight 2.183
    item osd.306 weight 2.183
    item osd.307 weight 2.183
    item osd.308 weight 2.183
    item osd.309 weight 2.183
    item osd.310 weight 2.183
    item osd.311 weight 2.183
    item osd.312 weight 2.183
    item osd.313 weight 2.183
    item osd.314 weight 2.183
}
root default {
    id -1        # do not change unnecessarily
    id -2 class hdd        # do not change unnecessarily
    id -51 class ssd        # do not change unnecessarily
    # weight 523.920
    alg straw2
    hash 0    # rjenkins1
    item dt-1ap213-proxmox-01 weight 32.745
    item dt-1ap213-proxmox-02 weight 32.745
    item dt-1ap213-proxmox-03 weight 32.745
    item dt-1ap213-proxmox-04 weight 32.745
    item dt-1ap213-proxmox-05 weight 32.745
    item dt-1ap213-proxmox-06 weight 32.745
    item dt-1ap213-proxmox-07 weight 32.745
    item dt-1ap213-proxmox-08 weight 32.745
    item dt-1ap214-proxmox-01 weight 32.745
    item dt-1ap214-proxmox-02 weight 32.745
    item dt-1ap214-proxmox-03 weight 32.745
    item dt-1ap214-proxmox-04 weight 32.745
    item dt-1ap214-proxmox-05 weight 32.745
    item dt-1ap214-proxmox-06 weight 32.745
    item dt-1ap214-proxmox-07 weight 32.745
    item dt-1ap214-proxmox-08 weight 32.745
}
host dt-1ap215-proxmox-01 {
    id -52        # do not change unnecessarily
    id -53 class hdd        # do not change unnecessarily
    id -54 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.320 weight 0.873
    item osd.321 weight 0.873
    item osd.322 weight 0.873
    item osd.323 weight 0.873
    item osd.324 weight 0.873
    item osd.325 weight 0.873
    item osd.326 weight 0.873
    item osd.327 weight 0.873
    item osd.328 weight 0.873
    item osd.329 weight 0.873
    item osd.330 weight 0.873
    item osd.331 weight 0.873
}
host dt-1ap215-proxmox-02 {
    id -55        # do not change unnecessarily
    id -56 class hdd        # do not change unnecessarily
    id -57 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.332 weight 0.873
    item osd.333 weight 0.873
    item osd.334 weight 0.873
    item osd.335 weight 0.873
    item osd.336 weight 0.873
    item osd.337 weight 0.873
    item osd.338 weight 0.873
    item osd.339 weight 0.873
    item osd.340 weight 0.873
    item osd.341 weight 0.873
    item osd.342 weight 0.873
    item osd.343 weight 0.873
}
host dt-1ap215-proxmox-03 {
    id -58        # do not change unnecessarily
    id -59 class hdd        # do not change unnecessarily
    id -60 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.344 weight 0.873
    item osd.345 weight 0.873
    item osd.346 weight 0.873
    item osd.347 weight 0.873
    item osd.348 weight 0.873
    item osd.349 weight 0.873
    item osd.350 weight 0.873
    item osd.351 weight 0.873
    item osd.352 weight 0.873
    item osd.353 weight 0.873
    item osd.354 weight 0.873
    item osd.355 weight 0.873
}
host dt-1ap215-proxmox-04 {
    id -61        # do not change unnecessarily
    id -62 class hdd        # do not change unnecessarily
    id -63 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.356 weight 0.873
    item osd.357 weight 0.873
    item osd.358 weight 0.873
    item osd.359 weight 0.873
    item osd.360 weight 0.873
    item osd.361 weight 0.873
    item osd.362 weight 0.873
    item osd.363 weight 0.873
    item osd.364 weight 0.873
    item osd.365 weight 0.873
    item osd.366 weight 0.873
    item osd.367 weight 0.873
}
host dt-1ap215-proxmox-05 {
    id -64        # do not change unnecessarily
    id -65 class hdd        # do not change unnecessarily
    id -66 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.368 weight 0.873
    item osd.369 weight 0.873
    item osd.370 weight 0.873
    item osd.371 weight 0.873
    item osd.373 weight 0.873
    item osd.372 weight 0.873
    item osd.374 weight 0.873
    item osd.375 weight 0.873
    item osd.376 weight 0.873
    item osd.377 weight 0.873
    item osd.378 weight 0.873
    item osd.379 weight 0.873
}
host dt-1ap215-proxmox-06 {
    id -67        # do not change unnecessarily
    id -68 class hdd        # do not change unnecessarily
    id -69 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.380 weight 0.873
    item osd.381 weight 0.873
    item osd.382 weight 0.873
    item osd.383 weight 0.873
    item osd.384 weight 0.873
    item osd.385 weight 0.873
    item osd.386 weight 0.873
    item osd.387 weight 0.873
    item osd.388 weight 0.873
    item osd.389 weight 0.873
    item osd.390 weight 0.873
    item osd.391 weight 0.873
}
host dt-1ap215-proxmox-07 {
    id -70        # do not change unnecessarily
    id -71 class hdd        # do not change unnecessarily
    id -72 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.392 weight 0.873
    item osd.393 weight 0.873
    item osd.394 weight 0.873
    item osd.395 weight 0.873
    item osd.396 weight 0.873
    item osd.397 weight 0.873
    item osd.398 weight 0.873
    item osd.399 weight 0.873
    item osd.400 weight 0.873
    item osd.401 weight 0.873
    item osd.402 weight 0.873
    item osd.403 weight 0.873
}
host dt-1ap215-proxmox-08 {
    id -73        # do not change unnecessarily
    id -74 class hdd        # do not change unnecessarily
    id -75 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.404 weight 0.873
    item osd.405 weight 0.873
    item osd.406 weight 0.873
    item osd.407 weight 0.873
    item osd.408 weight 0.873
    item osd.409 weight 0.873
    item osd.410 weight 0.873
    item osd.411 weight 0.873
    item osd.412 weight 0.873
    item osd.413 weight 0.873
    item osd.414 weight 0.873
    item osd.415 weight 0.873
}
host dt-1ap216-proxmox-05 {
    id -76        # do not change unnecessarily
    id -77 class hdd        # do not change unnecessarily
    id -78 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.416 weight 0.873
    item osd.417 weight 0.873
    item osd.418 weight 0.873
    item osd.419 weight 0.873
    item osd.420 weight 0.873
    item osd.421 weight 0.873
    item osd.422 weight 0.873
    item osd.423 weight 0.873
    item osd.424 weight 0.873
    item osd.425 weight 0.873
    item osd.426 weight 0.873
    item osd.427 weight 0.873
}
host dt-1ap216-proxmox-06 {
    id -79        # do not change unnecessarily
    id -80 class hdd        # do not change unnecessarily
    id -81 class ssd        # do not change unnecessarily
    # weight 10.476
    alg straw2
    hash 0    # rjenkins1
    item osd.428 weight 0.873
    item osd.429 weight 0.873
    item osd.430 weight 0.873
    item osd.431 weight 0.873
    item osd.432 weight 0.873
    item osd.433 weight 0.873
    item osd.434 weight 0.873
    item osd.435 weight 0.873
    item osd.436 weight 0.873
    item osd.437 weight 0.873
    item osd.438 weight 0.873
    item osd.439 weight 0.873
}
root ssd {
    id -82        # do not change unnecessarily
    id -83 class hdd        # do not change unnecessarily
    id -84 class ssd        # do not change unnecessarily
    # weight 104.780
    alg straw2
    hash 0    # rjenkins1
    item dt-1ap215-proxmox-01 weight 10.478
    item dt-1ap215-proxmox-02 weight 10.478
    item dt-1ap215-proxmox-03 weight 10.478
    item dt-1ap215-proxmox-04 weight 10.478
    item dt-1ap215-proxmox-05 weight 10.478
    item dt-1ap215-proxmox-06 weight 10.478
    item dt-1ap215-proxmox-07 weight 10.478
    item dt-1ap215-proxmox-08 weight 10.478
    item dt-1ap216-proxmox-05 weight 10.478
    item dt-1ap216-proxmox-06 weight 10.478
}

# rules
rule replicated_rule {
    id 0
    type replicated
    min_size 1
    max_size 10
    step take default
    step chooseleaf firstn 0 type host
    step emit
}
rule ssd_replicated_rule {
    id 1
    type replicated
    min_size 1
    max_size 10
    step take ssd
    step chooseleaf firstn 0 type host
    step emit
}
View Code

 

posted @ 2022-10-12 00:05  cyh00001  阅读(574)  评论(0编辑  收藏  举报