ceph 集羣故障恢復

集羣規劃配置

master1    172.16.230.21
master2    172.16.230.22
master3  172.16.230.23
node1  172.16.230.26
node2  172.16.230.27
node3  172.16.23028node

 

1、 模擬monitor 宕機狀態

2. 測試刪除monitor節點, 把master3 關機
測試

 

 發現 master3 節點已經宕機, 具體操做步驟,須要刪除配置文件中master3 信息,同步ceph.conf 配置文件 ,而後命令刪除master3

3. 修改ceph.conf 配置文件,刪除 monitor3信息ui

[root@master1 cluster-ceph]# cd /opt/cluster-ceph/

[global]
fsid = 574c4cb4-50f8-4d80-a61e-25eadd0c567d
mon_initial_members = master1, master2
mon_host = 172.16.230.21,172.16.230.22
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
public_network = 172.16.230.0/24
osd_pool_default_size = 2
mon_pg_warn_max_per_osd = 1000
osd pool default pg num = 256
osd pool default pgp num = 256
mon clock drift allowed = 2
mon clock drift warn backoff = 30

# 刪除 mon_initial_members  中的master3 和 mon_host 中 172.16.230.23 

 

4 . ceph.conf 同步到其餘節點spa

 ceph-deploy  --overwrite-conf  admin master1  master2  node1 node2 node3

 

 5. 使用remove命令 刪除節點code

[root@master1 cluster-ceph]# ceph mon remove master3
removing mon.master3 at 172.16.230.23:6789/0, there will be 2 monitors

 

6.  查看ceph集羣狀態 blog

 

 二.添加monitor3 到ceph集羣中(ceph-deploy)rem

[root@master1 cluster-ceph]# cd /opt/cluster-ceph/
[root@master1 cluster-ceph]# ceph-deploy mon create master3

 

同步ceph.conf 到集羣各個節點同步

 ceph-deploy  --overwrite-conf  admin master1  master2  node1 node2 node3

 

查看集羣狀態it

 

 

 

 

 

 

 

 

 

參考 https://www.bookstack.cn/read/ceph-handbook/Operation-add_rm_mon.mdio

相關文章
相關標籤/搜索