Ceph:删除OSD

Posted

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了Ceph:删除OSD相关的知识,希望对你有一定的参考价值。

1、查看OSD的分布信息:

# ceph osd tree
ID WEIGHT  TYPE NAME                     UP/DOWN REWEIGHT PRIMARY-AFFINITY 
-1 0.21950 root default                                                    
-2 0.04390     host hz-01-ops-tc-ceph-01                                   
 0 0.04390         osd.0                      up  1.00000          1.00000 
-3 0.04390     host hz-01-ops-tc-ceph-02                                   
 1 0.04390         osd.1                      up  1.00000          1.00000 
-4 0.04390     host hz-01-ops-tc-ceph-03                                   
 2 0.04390         osd.2                      up  1.00000          1.00000 
-5 0.04390     host hz-01-ops-tc-ceph-04                                   
 3 0.04390         osd.3                      up  1.00000          1.00000 
-6 0.04390     host hz01-dev-ops-wanl-01                                   
 4 0.04390         osd.4                      up  1.00000          1.00000

2、将osd.4移除OSD

# ceph osd out 4
marked out osd.4. 
# ceph -s
    cluster e2ca994a-00c4-477f-9390-ea3f931c5062
     health HEALTH_WARN
            56 pgs degraded
            1 pgs recovering
            55 pgs recovery_wait
            56 pgs stuck unclean
            recovery 604/1692 objects degraded (35.697%)
     monmap e1: 3 mons at {hz-01-ops-tc-ceph-02=172.16.2.231:6789/0,hz-01-ops-tc-ceph-03=172.16.2.172:6789/0,hz-01-ops-tc-ceph-04=172.16.2.181:6789/0}
            election epoch 20, quorum 0,1,2 hz-01-ops-tc-ceph-03,hz-01-ops-tc-ceph-04,hz-01-ops-tc-ceph-02
     osdmap e89: 5 osds: 5 up, 4 in
            flags sortbitwise,require_jewel_osds
      pgmap v68654: 1172 pgs, 4 pools, 2159 MB data, 564 objects
            5491 MB used, 174 GB / 179 GB avail
            604/1692 objects degraded (35.697%)
                1116 active+clean
                  55 active+recovery_wait+degraded
                   1 active+recovering+degraded
recovery io 87376 kB/s, 1 keys/s, 23 objects/s

3、在被删除OSD的那台ceph节点上的操作:

# systemctl stop [email protected]
# ceph osd tree
ID WEIGHT  TYPE NAME                     UP/DOWN REWEIGHT PRIMARY-AFFINITY 
-1 0.21950 root default                                                    
-2 0.04390     host hz-01-ops-tc-ceph-01                                   
 0 0.04390         osd.0                      up  1.00000          1.00000 
-3 0.04390     host hz-01-ops-tc-ceph-02                                   
 1 0.04390         osd.1                      up  1.00000          1.00000 
-4 0.04390     host hz-01-ops-tc-ceph-03                                   
 2 0.04390         osd.2                      up  1.00000          1.00000 
-5 0.04390     host hz-01-ops-tc-ceph-04                                   
 3 0.04390         osd.3                      up  1.00000          1.00000 
-6 0.04390     host hz01-dev-ops-wanl-01                                   
 4 0.04390         osd.4                    down        0          1.00000 
# ceph osd crush remove osd.4
removed item id 4 name 'osd.4' from crush map
# ceph auth del osd.4
updated
# ceph osd tree
ID WEIGHT  TYPE NAME                     UP/DOWN REWEIGHT PRIMARY-AFFINITY 
-1 0.17560 root default                                                    
-2 0.04390     host hz-01-ops-tc-ceph-01                                   
 0 0.04390         osd.0                      up  1.00000          1.00000 
-3 0.04390     host hz-01-ops-tc-ceph-02                                   
 1 0.04390         osd.1                      up  1.00000          1.00000 
-4 0.04390     host hz-01-ops-tc-ceph-03                                   
 2 0.04390         osd.2                      up  1.00000          1.00000 
-5 0.04390     host hz-01-ops-tc-ceph-04                                   
 3 0.04390         osd.3                      up  1.00000          1.00000 
-6       0     host hz01-dev-ops-wanl-01                                   
 4       0 osd.4                            down        0          1.00000 
# df -h
Filesystem           Size  Used Avail Use% Mounted on
/dev/mapper/cl-root   39G  1.8G   37G   5% /
devtmpfs             486M     0  486M   0% /dev
tmpfs                497M   84K  497M   1% /dev/shm
tmpfs                497M   26M  472M   6% /run
tmpfs                497M     0  497M   0% /sys/fs/cgroup
/dev/vda1           1014M  121M  894M  12% /boot
/dev/mapper/cl-home   19G   33M   19G   1% /home
/dev/vdb1             45G  237M   45G   1% /var/lib/ceph/osd/ceph-4
tmpfs                100M     0  100M   0% /run/user/0
# lsblk
NAME        MAJ:MIN RM  SIZE RO TYPE MOUNTPOINT
sr0          11:0    1 1024M  0 rom  
vda         252:0    0   60G  0 disk 
├─vda1      252:1    0    1G  0 part /boot
└─vda2      252:2    0   59G  0 part 
  ├─cl-root 253:0    0 38.3G  0 lvm  /
  ├─cl-swap 253:1    0    2G  0 lvm  [SWAP]
  └─cl-home 253:2    0 18.7G  0 lvm  /home
vdb         252:16   0   50G  0 disk 
├─vdb1      252:17   0   45G  0 part /var/lib/ceph/osd/ceph-4
└─vdb2      252:18   0    5G  0 part 
# ceph -s
    cluster e2ca994a-00c4-477f-9390-ea3f931c5062
     health HEALTH_OK
     monmap e1: 3 mons at {hz-01-ops-tc-ceph-02=172.16.2.231:6789/0,hz-01-ops-tc-ceph-03=172.16.2.172:6789/0,hz-01-ops-tc-ceph-04=172.16.2.181:6789/0}
            election epoch 20, quorum 0,1,2 hz-01-ops-tc-ceph-03,hz-01-ops-tc-ceph-04,hz-01-ops-tc-ceph-02
     osdmap e79: 4 osds: 4 up, 4 in
            flags sortbitwise,require_jewel_osds
      pgmap v68497: 1164 pgs, 3 pools, 2159 MB data, 564 objects
            6655 MB used, 173 GB / 179 GB avail
                1164 active+clean


以上是关于Ceph:删除OSD的主要内容,如果未能解决你的问题,请参考以下文章

ceph 创建和删除osd

ceph 删除osd的一些问题

ceph 删除了osd但是osd目录保存完整如何恢复

Ceph:删除OSD

ceph修复osd为down的情况

ceph磁盘管理添加删除