因为实验用的是虚拟机的关系,晚上一般会挂起。
第二天早上都能看到 4 slow ops, oldest one blocked for 638 sec, mon.cephnode01 has slow ops的报错。虽然不影响使用,但看着有点难受。
[root@ansible-01 ~]# ceph -s
cluster:
id: 38d10e08-7047-4341-bc8c-dbae13605cd6
health: HEALTH_WARN
4 slow ops, oldest one blocked for 638 sec, mon.cephnode01 has slow ops
services:
mon: 3 daemons, quorum cephnode01,cephnode02,cephnode03 (age 41m)
mgr: cephnode01(active, since 44h), standbys: cephnode03, cephnode02
mds: cephfs:1 {
0=cephnode01=up:active} 2 up:standby
osd: 6 osds: 6 up (since 25h), 6 in (since 2d)
rgw: 1 daemon active (cephnode01)
data:
pools: 10 pools, 144 pgs
objects: 276 objects, 29 MiB
usage: 6.2 GiB used, 18 GiB / 24 GiB avail
pgs: 144 active+clean
三台Ceph服务器上同步时间,并重启服务。
[root@cephnode01 ~]# ntpdate time.windows.com
[root@cephnode01 ~]# systemctl restart ceph-mon.target
看到HEALTH_OK,就很舒服了
[root@ansible-01 ~]# ceph health detail
HEALTH_OK
[root@ansible-01 ~]# ceph -s
cluster:
id: 38d10e08-7047-4341-bc8c-dbae13605cd6
health: HEALTH_OK
services:
mon: 3 daemons, quorum cephnode01,cephnode02,cephnode03 (age 74s)
mgr: cephnode01(active, since 45h), standbys: cephnode03, cephnode02
mds: cephfs:1 {
0=cephnode01=up:active} 2 up:standby
osd: 6 osds: 6 up (since 26h), 6 in (since 2d)
rgw: 1 daemon active (cephnode01)
data:
pools: 10 pools, 144 pgs
objects: 276 objects, 29 MiB
usage: 6.2 GiB used, 18 GiB / 24 GiB avail
pgs: 144 active+clean