Copy to ClipboardCopied!Toggle word wrapToggle overflow
SSH 到 ceph-0,然后
sudo “watch podman ps” # watch the new mon/mgr being deployed here
sudo “watch podman ps” # watch the new mon/mgr being deployed here
Copy to ClipboardCopied!Toggle word wrapToggle overflow
(可选)如果 mgr 在源节点上处于活跃状态,则:
ceph mgr fail <mgr instance>
ceph mgr fail <mgr instance>
Copy to ClipboardCopied!Toggle word wrapToggle overflow
从 cephadm shell 中,删除 oc0-controller-1 上的标签
for label in mon mgr _admin; do
ceph orch host rm label oc0-controller-1 $label;
done
for label in mon mgr _admin; do
ceph orch host rm label oc0-controller-1 $label;
done
Copy to ClipboardCopied!Toggle word wrapToggle overflow
将缺少的标签添加到 oc0-ceph-0
[ceph: root@oc0-controller-0 /]#
> for label in mon mgr _admin; do ceph orch host label add oc0-ceph-0 $label; done
Added label mon to host oc0-ceph-0
Added label mgr to host oc0-ceph-0
Added label _admin to host oc0-ceph-0
[ceph: root@oc0-controller-0 /]#
> for label in mon mgr _admin; do ceph orch host label add oc0-ceph-0 $label; done
Added label mon to host oc0-ceph-0
Added label mgr to host oc0-ceph-0
Added label _admin to host oc0-ceph-0
Copy to ClipboardCopied!Toggle word wrapToggle overflow
drain 和 force-remove oc0-controller-1 节点
[ceph: root@oc0-controller-0 /]# ceph orch host drain oc0-controller-1
Scheduled to remove the following daemons from host 'oc0-controller-1'
type id
-------------------- ---------------
mon oc0-controller-1
mgr oc0-controller-1.mtxohd
crash oc0-controller-1
[ceph: root@oc0-controller-0 /]# ceph orch host drain oc0-controller-1
Scheduled to remove the following daemons from host 'oc0-controller-1'
type id
-------------------- ---------------
mon oc0-controller-1
mgr oc0-controller-1.mtxohd
crash oc0-controller-1
Copy to ClipboardCopied!Toggle word wrapToggle overflow
[root@oc0-controller-1 ~]# sudo podman ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
5c1ad36472bc registry.redhat.io/ceph/rhceph@sha256:320c364dcc8fc8120e2a42f54eb39ecdba12401a2546763b7bef15b02ce93bc4 -n mon.oc0-contro... 35 minutes ago Up 35 minutes ago ceph-f6ec3ebe-26f7-56c8-985d-eb974e8e08e3-mon-oc0-controller-1
3b14cc7bf4dd registry.redhat.io/ceph/rhceph@sha256:320c364dcc8fc8120e2a42f54eb39ecdba12401a2546763b7bef15b02ce93bc4 -n mgr.oc0-contro... 35 minutes ago Up 35 minutes ago ceph-f6ec3ebe-26f7-56c8-985d-eb974e8e08e3-mgr-oc0-controller-1-mtxohd
[root@oc0-controller-1 ~]# cephadm rm-cluster --fsid f6ec3ebe-26f7-56c8-985d-eb974e8e08e3 --force
[root@oc0-controller-1 ~]# sudo podman ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
Copy to ClipboardCopied!Toggle word wrapToggle overflow
Copy to ClipboardCopied!Toggle word wrapToggle overflow
检查 oc0-ceph-0 节点中的新容器:
b581dc8bbb78 registry.redhat.io/ceph/rhceph@sha256:320c364dcc8fc8120e2a42f54eb39ecdba12401a2546763b7bef15b02ce93bc4 -n mon.oc0-ceph-0... 24 seconds ago Up 24 seconds ago ceph-f6ec3ebe-26f7-56c8-985d-eb974e8e08e3-mon-oc0-ceph-0
b581dc8bbb78 registry.redhat.io/ceph/rhceph@sha256:320c364dcc8fc8120e2a42f54eb39ecdba12401a2546763b7bef15b02ce93bc4 -n mon.oc0-ceph-0... 24 seconds ago Up 24 seconds ago ceph-f6ec3ebe-26f7-56c8-985d-eb974e8e08e3-mon-oc0-ceph-0
Copy to ClipboardCopied!Toggle word wrapToggle overflow
Copy to ClipboardCopied!Toggle word wrapToggle overflow
应用生成的 spec:
ceph orch apply -i ceph_spec.yaml
The result of 12 is having a new mgr deployed on the oc0-ceph-0 node, and the spec reconciled within cephadm
[ceph: root@oc0-controller-0 specs]# ceph orch ls
NAME PORTS RUNNING REFRESHED AGE PLACEMENT
crash 4/4 5m ago 61m *
mgr 3/3 5m ago 69s oc0-controller-0;oc0-ceph-0;oc0-controller-2
mon 3/3 5m ago 70s oc0-controller-0;oc0-ceph-0;oc0-controller-2
osd.default_drive_group 8 2m ago 69s oc0-ceph-0;oc0-ceph-1
[ceph: root@oc0-controller-0 specs]# ceph -s
cluster:
id: f6ec3ebe-26f7-56c8-985d-eb974e8e08e3
health: HEALTH_WARN
1 stray host(s) with 1 daemon(s) not managed by cephadm
services:
mon: 3 daemons, quorum oc0-controller-0,oc0-controller-2,oc0-ceph-0 (age 5m)
mgr: oc0-controller-0.xzgtvo(active, since 62m), standbys: oc0-controller-2.ahrgsk, oc0-ceph-0.hccsbb
osd: 8 osds: 8 up (since 42m), 8 in (since 49m); 1 remapped pgs
data:
pools: 1 pools, 1 pgs
objects: 0 objects, 0 B
usage: 43 MiB used, 400 GiB / 400 GiB avail
pgs: 1 active+clean
ceph orch apply -i ceph_spec.yaml
The result of 12 is having a new mgr deployed on the oc0-ceph-0 node, and the spec reconciled within cephadm
[ceph: root@oc0-controller-0 specs]# ceph orch ls
NAME PORTS RUNNING REFRESHED AGE PLACEMENT
crash 4/4 5m ago 61m *
mgr 3/3 5m ago 69s oc0-controller-0;oc0-ceph-0;oc0-controller-2
mon 3/3 5m ago 70s oc0-controller-0;oc0-ceph-0;oc0-controller-2
osd.default_drive_group 8 2m ago 69s oc0-ceph-0;oc0-ceph-1
[ceph: root@oc0-controller-0 specs]# ceph -s
cluster:
id: f6ec3ebe-26f7-56c8-985d-eb974e8e08e3
health: HEALTH_WARN
1 stray host(s) with 1 daemon(s) not managed by cephadm
services:
mon: 3 daemons, quorum oc0-controller-0,oc0-controller-2,oc0-ceph-0 (age 5m)
mgr: oc0-controller-0.xzgtvo(active, since 62m), standbys: oc0-controller-2.ahrgsk, oc0-ceph-0.hccsbb
osd: 8 osds: 8 up (since 42m), 8 in (since 49m); 1 remapped pgs
data:
pools: 1 pools, 1 pgs
objects: 0 objects, 0 B
usage: 43 MiB used, 400 GiB / 400 GiB avail
pgs: 1 active+clean
Copy to ClipboardCopied!Toggle word wrapToggle overflow
通过刷新 mgr 来修复警告:
ceph mgr fail oc0-controller-0.xzgtvo
ceph mgr fail oc0-controller-0.xzgtvo
Copy to ClipboardCopied!Toggle word wrapToggle overflow