基于cephadm部署ceph集群

一、环境准备

1.1、制作模板时将docker先安装

[root@ubuntu-20-04 docker]# pwd
/root/docker
[root@ubuntu-20-04 docker]# ll -h 
total 74M
drwxr-xr-x 3 root   root   4.0K Oct  5 14:21 ./
drwx------ 6 root   root   4.0K Oct  5 14:21 ../
-rw-r--r-- 1 root   root    647 Apr 11  2021 containerd.service
-rw-r--r-- 1 root   root    356 Jul 18 11:09 daemon.json
drwxrwxr-x 2 docker docker 4.0K Jun  6 23:03 docker/
-rw-r--r-- 1 root   root    62M Jun  7 08:42 docker-20.10.17.tgz
-rwxr-xr-x 1 root   root    12M Dec  7  2021 docker-compose-Linux-x86_64_1.28.6*
-rwxr-xr-x 1 root   root   2.9K Jul 21 07:29 docker-install.sh*
-rw-r--r-- 1 root   root   1.7K Apr 11  2021 docker.service
-rw-r--r-- 1 root   root    197 Apr 11  2021 docker.socket
-rw-r--r-- 1 root   root    454 Apr 11  2021 limits.conf
-rw-r--r-- 1 root   root    257 Apr 11  2021 sysctl.conf
[root@ubuntu-20-04 docker]# 

bash docker-install.sh

+ pwd
+ DIR=/root/docker
+ PACKAGE_NAME=docker-20.10.17.tgz
+ DOCKER_FILE=/root/docker/docker-20.10.17.tgz
+ main
+ centos_install_docker
+ 
+ [ 0 -eq 0 ]
+ grep Kernel+ cat /etc/redhat-release
 /etc/issue
cat: /etc/redhat-release: No such file or directory
+ /bin/echo 当前系统是,即将开始系统初始化、配置docker-compose与安装docker
当前系统是,即将开始系统初始化、配置docker-compose与安装docker
+ sleep 1
+ systemctl stop firewalld
Failed to stop firewalld.service: Unit firewalld.service not loaded.
+ systemctl stop NetworkManager
+ systemctl disable NetworkManager
Removed /etc/systemd/system/dbus-org.freedesktop.nm-dispatcher.service.
Removed /etc/systemd/system/multi-user.target.wants/NetworkManager.service.
Removed /etc/systemd/system/network-online.target.wants/NetworkManager-wait-online.service.
+ echo NetworkManager
NetworkManager
+ sleep 1
+ sed -i s/SELINUX=enforcing/SELINUX=disabled/g /etc/sysconfig/selinux
sed: can't read /etc/sysconfig/selinux: No such file or directory
+ cp /root/docker/limits.conf /etc/security/limits.conf
+ cp /root/docker/sysctl.conf /etc/sysctl.conf
+ /bin/tar xvf /root/docker/docker-20.10.17.tgz
docker/
docker/docker-init
docker/containerd
docker/ctr
docker/runc
docker/dockerd
docker/docker-proxy
docker/containerd-shim
docker/docker
docker/containerd-shim-runc-v2
+ cp docker/containerd docker/containerd-shim docker/containerd-shim-runc-v2 docker/ctr docker/docker docker/docker-init docker/docker-proxy docker/dockerd docker/runc /usr/bin
+ mkdir /etc/docker
+ cp daemon.json /etc/docker
+ cp containerd.service /lib/systemd/system/containerd.service
+ cp docker.service /lib/systemd/system/docker.service
+ cp docker.socket /lib/systemd/system/docker.socket
+ cp /root/docker/docker-compose-Linux-x86_64_1.28.6 /usr/bin/docker-compose
+ groupadd docker
+ useradd docker -s /sbin/nologin -g docker
+ 
+ [ 0 -ne 0 ]
+ usermod magedu -G docker
+ id -u magedu
id: ‘magedu’: no such user
usermod: user 'magedu' does not exist
+ systemctl enable containerd.service
Created symlink /etc/systemd/system/multi-user.target.wants/containerd.service → /lib/systemd/system/containerd.service.
+ systemctl restart containerd.service
+ systemctl enable docker.service
Created symlink /etc/systemd/system/multi-user.target.wants/docker.service → /lib/systemd/system/docker.service.
+ systemctl restart docker.service
+ systemctl enable docker.socket
Created symlink /etc/systemd/system/sockets.target.wants/docker.socket → /lib/systemd/system/docker.socket.
+ systemctl restart docker.socket
+ ubuntu_install_docker
+ 
+ + [ 0 -eqgrep 0 Ubuntu ] /etc/issue

+ cat /etc/issue
+ /bin/echo 当前系统是Ubuntu 20.04.4 LTS \n \l,即将开始系统初始化、配置docker-compose与安装docker
Ubuntu 20.04.4 LTS \n \l
当前系统是Ubuntu 20.04.4 LTS \n \l,即将开始系统初始化、配置docker-compose与安装docker
+ sleep 1
+ cp /root/docker/limits.conf /etc/security/limits.conf
+ cp /root/docker/sysctl.conf /etc/sysctl.conf
+ /bin/tar xvf /root/docker/docker-20.10.17.tgz
docker/
docker/docker-init
docker/containerd
docker/ctr
docker/runc
docker/dockerd
docker/docker-proxy
docker/containerd-shim
docker/docker
docker/containerd-shim-runc-v2
+ cp docker/containerd docker/containerd-shim docker/containerd-shim-runc-v2 docker/ctr docker/docker docker/docker-init docker/docker-proxy docker/dockerd docker/runc /usr/bin
cp: cannot create regular file '/usr/bin/containerd': Text file busy
cp: cannot create regular file '/usr/bin/dockerd': Text file busy
+ mkdir /etc/docker
mkdir: cannot create directory ‘/etc/docker’: File exists
+ cp containerd.service /lib/systemd/system/containerd.service
+ cp docker.service /lib/systemd/system/docker.service
+ cp docker.socket /lib/systemd/system/docker.socket
+ cp /root/docker/docker-compose-Linux-x86_64_1.28.6 /usr/bin/docker-compose
+ /bin/echo docker 安装完成!
docker 安装完成!
+ sleep 1
+ groupadd docker
groupadd: group 'docker' already exists
+ 
+ [ 0 -ne 0 ]
+ usermod magedu -G docker
+ id -u magedu
id: ‘magedu’: no such user
usermod: user 'magedu' does not exist
+ systemctl enable containerd.service
+ systemctl restart containerd.service
+ systemctl enable docker.service
+ systemctl restart docker.service
+ systemctl enable docker.socket
+ systemctl restart docker.socket

1.2、机器准备

172.16.88.131 cephadm-ceph-mon-01    2vcpu 4G 50G 
172.16.88.132 cephadm-ceph-mon-01    2vcpu 4G 50G 
172.16.88.133 cephadm-ceph-mon-01    2vcpu 4G 50G 
172.16.88.134 cephadm-ceph-mgr-01    2vcpu 4G 50G 
172.16.88.135 cephadm-ceph-mgr-02    2vcpu 4G 50G 
172.16.88.136 cephadm-ceph-node-01   2vcpu 4G 50G 3*50G
172.16.88.137 cephadm-ceph-node-02   2vcpu 4G 50G 3*50G
172.16.88.138 cephadm-ceph-node-03   2vcpu 4G 50G 3*50G

1.3、在mon-01节点安装cephadm

[root@cephadm-ceph-mon-01 ~]# wget https://mirrors.aliyun.com/ceph/debian-17.2.4/pool/main/c/ceph/cephadm_17.2.4-1focal_amd64.deb

[root@cephadm-ceph-mon-01 ~]# dpkg -i cephadm_17.2.4-1focal_amd64.deb

[root@cephadm-ceph-mon-01 ~]# cephadm pull  ##所有机器都提前下载可以使用docker pull quay.io/ceph/ceph:v17

[root@cephadm-ceph-mon-01 ~]# cephadm pull
Pulling container image quay.io/ceph/ceph:v17...
{
    "ceph_version": "ceph version 17.2.4 (1353ed37dec8d74973edc3d5d5908c20ad5a7332) quincy (stable)",
    "image_id": "652a1437d122ef2016e6a87c4653bd58691e9255e3efb8f31a3a4df3f82c0636",
    "repo_digests": [
        "quay.io/ceph/ceph@sha256:ffc55676b680c2428f7feb2daa331050d8606e2cb43a86139dbb9a2e3b02488e"
    ]
}
[root@cephadm-ceph-mon-01 ~]# docker images
REPOSITORY          TAG       IMAGE ID       CREATED      SIZE
quay.io/ceph/ceph   v17       652a1437d122   4 days ago   1.34GB
[root@cephadm-ceph-mon-01 ~]# 

1.4、初始化 ceph 集群

当前节点安装 mon、 mgr 角色,部署 prometheus、 grafana、 alertmanager、 node-exporter等服务。

[root@cephadm-ceph-mon-01 ~]# cephadm bootstrap --mon-ip 172.16.88.131 --cluster-network 172.16.88.0/24 --allow-fqdn-hostname

Creating directory /etc/ceph for ceph.conf
Verifying podman|docker is present...
Verifying lvm2 is present...
Verifying time synchronization is in place...
Unit systemd-timesyncd.service is enabled and running
Repeating the final host check...
docker (/usr/bin/docker) is present
systemctl is present
lvcreate is present
Unit systemd-timesyncd.service is enabled and running
Host looks OK
Cluster fsid: 4d635f30-44c3-11ed-9895-1dd8970b38a5
Verifying IP 172.16.88.131 port 3300 ...
Verifying IP 172.16.88.131 port 6789 ...
Mon IP `172.16.88.131` is in CIDR network `172.16.88.0/24`
Mon IP `172.16.88.131` is in CIDR network `172.16.88.0/24`
Pulling container image quay.io/ceph/ceph:v17...
Ceph version: ceph version 17.2.4 (1353ed37dec8d74973edc3d5d5908c20ad5a7332) quincy (stable)
Extracting ceph user uid/gid from container image...
Creating initial keys...
Creating initial monmap...
Creating mon...
Waiting for mon to start...
Waiting for mon...
mon is available
Assimilating anything we can from ceph.conf...
Generating new minimal ceph.conf...
Restarting the monitor...
Setting mon public_network to 172.16.88.0/24
Setting cluster_network to 172.16.88.0/24
Wrote config to /etc/ceph/ceph.conf
Wrote keyring to /etc/ceph/ceph.client.admin.keyring
Creating mgr...
Verifying port 9283 ...
Waiting for mgr to start...
Waiting for mgr...
mgr not available, waiting (1/15)...
mgr not available, waiting (2/15)...
mgr not available, waiting (3/15)...
mgr not available, waiting (4/15)...
mgr not available, waiting (5/15)...
mgr is available
Enabling cephadm module...
Waiting for the mgr to restart...
Waiting for mgr epoch 5...
mgr epoch 5 is available
Setting orchestrator backend to cephadm...
Generating ssh key...
Wrote public SSH key to /etc/ceph/ceph.pub
Adding key to root@localhost authorized_keys...
Adding host cephadm-ceph-mon-01...
Deploying mon service with default placement...
Deploying mgr service with default placement...
Deploying crash service with default placement...
Deploying prometheus service with default placement...
Deploying grafana service with default placement...
Deploying node-exporter service with default placement...
Deploying alertmanager service with default placement...
Enabling the dashboard module...
Waiting for the mgr to restart...
Waiting for mgr epoch 9...
mgr epoch 9 is available
Generating a dashboard self-signed certificate...
Creating initial admin user...
Fetching dashboard port number...
Ceph Dashboard is now available at:

         URL: https://cephadm-ceph-mon-01:8443/
        User: admin
    Password: l59dve9gx6

Enabling client.admin keyring and conf on hosts with "admin" label
Saving cluster configuration to /var/lib/ceph/4d635f30-44c3-11ed-9895-1dd8970b38a5/config directory
Enabling autotune for osd_memory_target
You can access the Ceph CLI as following in case of multi-cluster or non-default config:

    sudo /usr/sbin/cephadm shell --fsid 4d635f30-44c3-11ed-9895-1dd8970b38a5 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring

Or, if you are only running a single cluster on this host:

    sudo /usr/sbin/cephadm shell 

Please consider enabling telemetry to help improve Ceph:

    ceph telemetry on

For more information see:

    https://docs.ceph.com/docs/master/mgr/telemetry/

Bootstrap complete.

1.5、进入ceph管理终端

[root@cephadm-ceph-mon-01 ~]# sudo /usr/sbin/cephadm shell --fsid 4d635f30-44c3-11ed-9895-1dd8970b38a5 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring

安装ceph客户端命令

[root@cephadm-ceph-mon-01 ~]# cephadm install ceph-common

二、添加node节点

2.1、将node主机添加到集群

创建密钥,并分发到其他机器

[root@cephadm-ceph-mon-01 ~]# ssh-keygen 
Generating public/private rsa key pair.
Enter file in which to save the key (/root/.ssh/id_rsa): 
Enter passphrase (empty for no passphrase): 
Enter same passphrase again: 
Your identification has been saved in /root/.ssh/id_rsa
Your public key has been saved in /root/.ssh/id_rsa.pub
The key fingerprint is:
SHA256:ytOO40r2/oMZYkD60/y2P9yMkmziWFAQUJ0d/2gN5cc root@cephadm-ceph-mon-01
The key's randomart image is:
+---[RSA 3072]----+
|.o+o o..  .      |
|  ..o .. o .     |
| o  .   o . E    |
|. ..     = .     |
| ..+    S o      |
|  o.=..+         |
|   o+++*.+       |
|   =..@+= o      |
|  ..oB**+o       |
+----[SHA256]-----+
[root@cephadm-ceph-mon-01 ~]# 

[root@cephadm-ceph-mon-01 ~]# apt-get install sshpass

[root@cephadm-ceph-mon-01 ~]# vi ssh-key.sh 

#!/bin/bash
for i in {131..138};
do
      sshpass -p 'redhat' ssh-copy-id -o StrictHostKeyChecking=no -i /root/.ssh/id_rsa -p 22 root@172.16.88.$i;
done

[root@cephadm-ceph-mon-01 ~]# bash ssh-key.sh 

分发ceph.pub到其他机器

[root@cephadm-ceph-mon-01 ceph]# for i in {132..138};do ssh-copy-id -f -i /etc/ceph/ceph.pub root@172.16.88.$i;done

2.2、添加ceph-mon节点

[root@cephadm-ceph-mon-01 ~]# cephadm shell ceph orch host add cephadm-ceph-mon-02 172.16.88.132

Inferring fsid 4d635f30-44c3-11ed-9895-1dd8970b38a5
Inferring config /var/lib/ceph/4d635f30-44c3-11ed-9895-1dd8970b38a5/mon.cephadm-ceph-mon-01/config
Using ceph image with id '652a1437d122' and tag 'v17' created on 2022-10-01 00:31:25 +0800 CST
quay.io/ceph/ceph@sha256:ffc55676b680c2428f7feb2daa331050d8606e2cb43a86139dbb9a2e3b02488e
Added host 'cephadm-ceph-mon-02' with addr '172.16.88.132'

继续添加ceph-mon-03、ceph-node-01\02\03

cephadm shell ceph orch host add cephadm-ceph-mon-03 172.16.88.133
cephadm shell ceph orch host add cephadm-ceph-node-01 172.16.88.136
cephadm shell ceph orch host add cephadm-ceph-node-02 172.16.88.137
cephadm shell ceph orch host add cephadm-ceph-node-03 172.16.88.138

添加到集群的主机, 默认会部署 mon 服务

2.3、将磁盘添加到ceph集群

查看可用的磁盘

[root@cephadm-ceph-mon-01 ~]# cephadm shell ceph orch device ls

将所有node节点磁盘添加到ceph集群

ceph orch daemon add osd cephadm-ceph-node-01:/dev/vdb
ceph orch daemon add osd cephadm-ceph-node-02:/dev/vdc
ceph orch daemon add osd cephadm-ceph-node-03:/dev/vdd

ceph orch daemon add osd cephadm-ceph-node-01:/dev/vdb
ceph orch daemon add osd cephadm-ceph-node-02:/dev/vdc
ceph orch daemon add osd cephadm-ceph-node-03:/dev/vdd

ceph orch daemon add osd cephadm-ceph-node-01:/dev/vdb
ceph orch daemon add osd cephadm-ceph-node-02:/dev/vdc
ceph orch daemon add osd cephadm-ceph-node-03:/dev/vdd

验证集群状态

2.4、添加专用的ceph-mgr节点

[root@cephadm-ceph-mon-01 ~]# cephadm shell ceph orch host add cephadm-ceph-mgr-01 172.16.88.134

[root@cephadm-ceph-mon-01 ~]# cephadm shell ceph orch host add cephadm-ceph-mgr-01 172.16.88.134
Inferring fsid 4d635f30-44c3-11ed-9895-1dd8970b38a5
Inferring config /var/lib/ceph/4d635f30-44c3-11ed-9895-1dd8970b38a5/mon.cephadm-ceph-mon-01/config
Using ceph image with id '652a1437d122' and tag 'v17' created on 2022-10-01 00:31:25 +0800 CST
quay.io/ceph/ceph@sha256:ffc55676b680c2428f7feb2daa331050d8606e2cb43a86139dbb9a2e3b02488e
Added host 'cephadm-ceph-mgr-01' with addr '172.16.88.134'
[root@cephadm-ceph-mon-01 ~]# cephadm shell ceph orch host add cephadm-ceph-mgr-02 172.16.88.135
Inferring fsid 4d635f30-44c3-11ed-9895-1dd8970b38a5
Inferring config /var/lib/ceph/4d635f30-44c3-11ed-9895-1dd8970b38a5/mon.cephadm-ceph-mon-01/config
Using ceph image with id '652a1437d122' and tag 'v17' created on 2022-10-01 00:31:25 +0800 CST
quay.io/ceph/ceph@sha256:ffc55676b680c2428f7feb2daa331050d8606e2cb43a86139dbb9a2e3b02488e
Added host 'cephadm-ceph-mgr-02' with addr '172.16.88.135'
[root@cephadm-ceph-mon-01 ~]#

切换mgr节点

 [root@cephadm-ceph-mon-01 ~]# ceph orch apply mgr cephadm-ceph-mgr-01,cephadm-ceph-mgr-02

三、创建pool测试

[root@cephadm-ceph-mon-01 ~]# ceph osd pool create rbd-data1 32 32  #创建rbd-data1 pool池
pool 'rbd-data1' created
[root@cephadm-ceph-mon-01 ~]# 
[root@cephadm-ceph-mon-01 ~]# ceph osd pool ls #列出集群所有pool
.mgr
rbd-data1
[root@cephadm-ceph-mon-01 ~]# 
[root@cephadm-ceph-mon-01 ~]# ceph df
--- RAW STORAGE ---
CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED
hdd    450 GiB  450 GiB  190 MiB   190 MiB       0.04
TOTAL  450 GiB  450 GiB  190 MiB   190 MiB       0.04
 
--- POOLS ---
POOL       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL
.mgr        1    1  449 KiB        2  1.3 MiB      0    142 GiB
rbd-data1   2   32      0 B        0      0 B      0    142 GiB
[root@cephadm-ceph-mon-01 ~]# ceph osd pool application enable rbd-data1 rbd #启用rbd-data1池应用
enabled application 'rbd' on pool 'rbd-data1'
[root@cephadm-ceph-mon-01 ~]# rbd pool init -p rbd-data1
[root@cephadm-ceph-mon-01 ~]# rbd create data-img1 --size 3G --pool rbd-data1 --image-format 2 --image-feature layering #在rbd-data1池下创建3G大小data-img1镜像
[root@cephadm-ceph-mon-01 ~]# rbd create data-img2 --size 3G --pool rbd-data1 --image-format 2 --image-feature layering
[root@cephadm-ceph-mon-01 ~]# rbd ls --pool rbd-data1
data-img1
data-img2
[root@cephadm-ceph-mon-01 ~]# rbd ls --pool rbd-data1 -l
NAME       SIZE   PARENT  FMT  PROT  LOCK
data-img1  3 GiB            2            
data-img2  3 GiB            2            
[root@cephadm-ceph-mon-01 ~]# rbd --image data-img2 --pool rbd-data1 info #查看rbd-data1下data-img2镜像详细信息
rbd image 'data-img2':
    size 3 GiB in 768 objects
    order 22 (4 MiB objects)
    snapshot_count: 0
    id: 39c13b224fb9
    block_name_prefix: rbd_data.39c13b224fb9
    format: 2
    features: layering
    op_features: 
    flags: 
    create_timestamp: Thu Oct  6 21:20:59 2022
    access_timestamp: Thu Oct  6 21:20:59 2022
    modify_timestamp: Thu Oct  6 21:20:59 2022
[root@cephadm-ceph-mon-01 ~]# rbd --image data-img1 --pool rbd-data1 info
rbd image 'data-img1':
    size 3 GiB in 768 objects
    order 22 (4 MiB objects)
    snapshot_count: 0
    id: 39b7dc5e8e0a
    block_name_prefix: rbd_data.39b7dc5e8e0a
    format: 2
    features: layering
    op_features: 
    flags: 
    create_timestamp: Thu Oct  6 21:20:53 2022
    access_timestamp: Thu Oct  6 21:20:53 2022
    modify_timestamp: Thu Oct  6 21:20:53 2022
[root@cephadm-ceph-mon-01 ~]# 

挂载rbd

[root@cephadm-ceph-mon-01 ~]# rbd -p rbd-data1 map data-img1 #映射rbd到本地目录
/dev/rbd0
[root@cephadm-ceph-mon-01 ~]# lsblk
NAME                      MAJ:MIN RM  SIZE RO TYPE MOUNTPOINT
loop0                       7:0    0 67.2M  1 loop /snap/lxd/21835
loop1                       7:1    0   47M  1 loop /snap/snapd/16292
loop2                       7:2    0 67.8M  1 loop /snap/lxd/22753
loop3                       7:3    0   62M  1 loop /snap/core20/1587
loop4                       7:4    0 63.2M  1 loop /snap/core20/1623
loop5                       7:5    0   48M  1 loop /snap/snapd/17029
rbd0                      251:0    0    3G  0 disk 
vda                       252:0    0   50G  0 disk 
├─vda1                    252:1    0    1M  0 part 
├─vda2                    252:2    0  1.5G  0 part /boot
└─vda3                    252:3    0 48.5G  0 part 
  └─ubuntu--vg-ubuntu--lv 253:0    0 48.5G  0 lvm  /
[root@cephadm-ceph-mon-01 ~]# mkdir /rbd-data
[root@cephadm-ceph-mon-01 ~]# mkfs.xfs /dev/rbd0
meta-data=/dev/rbd0              isize=512    agcount=8, agsize=98304 blks
         =                       sectsz=512   attr=2, projid32bit=1
         =                       crc=1        finobt=1, sparse=1, rmapbt=0
         =                       reflink=1
data     =                       bsize=4096   blocks=786432, imaxpct=25
         =                       sunit=16     swidth=16 blks
naming   =version 2              bsize=4096   ascii-ci=0, ftype=1
log      =internal log           bsize=4096   blocks=2560, version=2
         =                       sectsz=512   sunit=16 blks, lazy-count=1
realtime =none                   extsz=4096   blocks=0, rtextents=0
[root@cephadm-ceph-mon-01 ~]# mount /dev/rbd0  /rbd-data/
[root@cephadm-ceph-mon-01 ~]# df -Th
Filesystem                        Type      Size  Used Avail Use% Mounted on
udev                              devtmpfs  1.9G     0  1.9G   0% /dev
tmpfs                             tmpfs     394M  1.7M  392M   1% /run
/dev/mapper/ubuntu--vg-ubuntu--lv ext4       48G   12G   35G  25% /
tmpfs                             tmpfs     2.0G     0  2.0G   0% /dev/shm
tmpfs                             tmpfs     5.0M     0  5.0M   0% /run/lock
tmpfs                             tmpfs     2.0G     0  2.0G   0% /sys/fs/cgroup
/dev/loop1                        squashfs   47M   47M     0 100% /snap/snapd/16292
/dev/loop0                        squashfs   68M   68M     0 100% /snap/lxd/21835
/dev/loop3                        squashfs   62M   62M     0 100% /snap/core20/1587
/dev/loop2                        squashfs   68M   68M     0 100% /snap/lxd/22753
/dev/vda2                         ext4      1.5G  205M  1.2G  15% /boot
/dev/loop5                        squashfs   48M   48M     0 100% /snap/snapd/17029
/dev/loop4                        squashfs   64M   64M     0 100% /snap/core20/1623
tmpfs                             tmpfs     394M     0  394M   0% /run/user/0
overlay                           overlay    48G   12G   35G  25% /var/lib/docker/overlay2/566d851e9c98bf7dc7fcfd9af57213657c3594981e9a90970dfcce54845b786c/merged
overlay                           overlay    48G   12G   35G  25% /var/lib/docker/overlay2/719c6850931528a0ccb13cd89f660a481f82923ac0d710874ed2fad8f58852e8/merged
overlay                           overlay    48G   12G   35G  25% /var/lib/docker/overlay2/3b2595d12e589f79a09ed36ea2e6760bfb73e04d866beb1352fe9230bc1e5520/merged
overlay                           overlay    48G   12G   35G  25% /var/lib/docker/overlay2/3f3e47fc108ba57d194a8dc55431e0c6d7bce1b4dddf2ffd4ac171193eafe63d/merged
overlay                           overlay    48G   12G   35G  25% /var/lib/docker/overlay2/53c9d28bf88152b45d148abdd80ed3a713a45c92c6f3c69452f7d8773a45928b/merged
overlay                           overlay    48G   12G   35G  25% /var/lib/docker/overlay2/8bf9de2d17ff69e1efc0463417f1d0ce5a4a85a56eea2c195f6470173b25f7e8/merged
/dev/rbd0                         xfs       3.0G   54M  3.0G   2% /rbd-data
[root@cephadm-ceph-mon-01 ~]# cd /rbd-data/
[root@cephadm-ceph-mon-01 rbd-data]# ls
[root@cephadm-ceph-mon-01 rbd-data]# touch test-{1..9}.txt
[root@cephadm-ceph-mon-01 rbd-data]# ll -h
total 4.0K
drwxr-xr-x  2 root root  168 Oct  6 21:29 ./
drwxr-xr-x 20 root root 4.0K Oct  6 21:27 ../
-rw-r--r--  1 root root    0 Oct  6 21:29 test-1.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-2.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-3.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-4.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-5.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-6.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-7.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-8.txt
-rw-r--r--  1 root root    0 Oct  6 21:29 test-9.txt
[root@cephadm-ceph-mon-01 rbd-data]# 

四、验证自带的Prometheus监控

修改grafana初始admin密码

[root@cephadm-ceph-mon-01 ~]# vim grafana.yaml 
[root@cephadm-ceph-mon-01 ~]# cat grafana.yaml 
service_type: grafana
spec:
  initial_admin_password: password12345678
[root@cephadm-ceph-mon-01 ~]#
[root@cephadm-ceph-mon-01 ~]# ceph orch  apply -i grafana.yaml 
Scheduled grafana update...
[root@cephadm-ceph-mon-01 ~]# ceph orch redeploy grafana
Scheduled to redeploy grafana.cephadm-ceph-mon-01 on host 'cephadm-ceph-mon-01'
[root@cephadm-ceph-mon-01 ~]# 
[root@cephadm-ceph-mon-01 ~]# 

 背景颜色调整

主机监控

pool池监控

 

posted @ 2022-10-06 09:00  cyh00001  阅读(2130)  评论(0编辑  收藏  举报