ceph/ceph-filestore.conf # must point to a valid ceph configuration file directory=/var/lib/ceph/osd/ceph -8/ # journal inside fio_dir osd journal = /var/lib/ceph/osd/ceph-8/journal osd journal -11-g42f1 Starting 1 process 2016-11-23 22:32:17.713473 7f7536d56780 0 filestore(/var/lib/ceph/osd/ceph config option 2016-11-23 22:32:17.805018 7f7536d56780 0 genericfilestorebackend(/var/lib/ceph/osd/ceph -8/) start omap initiation 2016-11-23 22:32:17.926589 7f7536d56780 0 filestore(/var/lib/ceph/osd/ceph
var/lib/ceph/osd/ceph-7 tmpfs 24700568 24 24700544 1% /var/lib/ceph/osd/ceph
数据基本查看其他正常的pg,对象数量平均是在6000多个 systemctl stop ceph-osd@8 ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph 11 删除两个osd节点上的故障pg副本 systemctl stop ceph-osd@8 ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph pgid 2.1c --op import --file /opt/2.1c.obj_osd.14 ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph
sdf : /dev/sdf1 ceph journal, for /dev/sde1 查看jounral的数据 [root@lab8106 ceph]# ll /var/lib/ceph/osd/ceph
'bgw-os-node153' from crush map 8)、卸载所有挂载在osd的硬盘 [root@bgw-os-node153 ~]# umount /var/lib/ceph/osd/ceph
ceph-volume lvm list ls /dev/mapper/ systemctl stop ceph-osd@8.service umount -l umount -l /var/lib/ceph/osd/ceph