ceph PG 状态恢复不了的问题 (来自ceph-devel邮件)
来源:互联网 发布:软件项目阶段性汇报 编辑:程序博客网 时间:2024/06/09 22:41
ceph health detail
HEALTH_WARN 2 pgs down; 2 pgs peering; 2 pgs stuck inactive
pg 1.165 is stuck inactive since forever, current state down+remapped+peering, last acting [38,48]
pg 1.60 is stuck inactive since forever, current state down+remapped+peering, last acting [66,40]
pg 1.60 is down+remapped+peering, acting [66,40]
pg 1.165 is down+remapped+peering, acting [38,48]
[root@cc1 ~]# ceph -s
cluster 8cdfbff9-b7be-46de-85bd-9d49866fcf60
health HEALTH_WARN
2 pgs down
2 pgs peering
2 pgs stuck inactive
monmap e1: 3 mons at {cc1=192.168.128.1:6789/0,cc2=192.168.128.2:6789/0,cc3=192.168.128.3:6789/0}
election epoch 872, quorum 0,1,2 cc1,cc2,cc3
osdmap e115175: 100 osds: 88 up, 86 in; 2 remapped pgs
pgmap v67583069: 3520 pgs, 17 pools, 26675 GB data, 4849 kobjects
76638 GB used, 107 TB / 182 TB avail
3515 active+clean
3 active+clean+scrubbing+deep
2 down+remapped+peering
client io 0 B/s rd, 869 kB/s wr, 14 op/s rd, 113 op/s wr
The thing where you can't query a PG is because the OSD is throttling
incoming work and the throttle is exhausted (the PG can't do work so it
isn't making progress). A workaround for jewel is to restart the OSD
serving the PG and do the query quickly after that (probably in a loop so
that you catch it after it starts up but before the throttle is
exhausted again). (In luminous this is fixed.)
Once you have the query output ('ceph tell $pgid query') you'll be able to
tell what is preventing the PG from peering.
You can identify the osd(s) hosting the pg with 'ceph pg map $pgid'.
If you haven't deleted the data, you should start the OSDs back up.
If they are partially damanged you can use ceph-objectstore-tool to
extract just the PGs in question to make sure you haven't lost anything,
inject them on some other OSD(s) and restart those, and *then* mark the
bad OSDs as 'lost'.
If all else fails, you can just mark those OSDs 'lost', but in doing so
you might be telling the cluster to lose data.
The best thing to do is definitely to get those OSDs started again.
阅读全文
0 0
- ceph PG 状态恢复不了的问题 (来自ceph-devel邮件)
- Ceph PG的不同状态
- ceph - pg 常见状态
- ceph: PG 状态
- ceph存储 PG的数据恢复过程
- Ceph PG 归置组状态
- ceph存储 ceph中PG的意义
- ceph中 PG的意义
- ceph pg(placement group)状态总结
- ceph pg和pgp的区别
- Ceph中PG和PGP的区别
- ceph pg split
- ceph pg数量调整
- ceph PG数 告警
- ceph学习之PG
- ceph 获取每个osd上pg的数量
- ceph存储 PG的状态机和peering过程
- ceph pg, osd, pool 之间的一些查找方法
- common-pool1.6连接池源码分析流程图
- 关于jar打包时的几个问题的汇总
- spring/springmvc/tomcat启动报错:Servlet mapping specifies an unknown servlet name taotao-search,Failed t
- Myeclipse 中SVN Could not create the view: org.tigris.subversion.subclipse.ui.repository.Repositories
- Oracle中Union与UnionAll的区别(union会合并多个重复结果)
- ceph PG 状态恢复不了的问题 (来自ceph-devel邮件)
- linux yum配置
- 【Linux】Linux中常用操作命令
- Linux Shell 自定义函数(定义、返回值、变量作用域)介绍
- JavaScript
- Protractor 环境搭建
- 【交通银行副行长 侯维栋】商业银行的科技发展趋势
- 创建双边偏置曲线(通过TOGGLE控件来控制是否保留原来的曲线)
- Calling Fortran function and subroutines from a C or C++ function