Jewel版cephfs当rm -rf文件时报错"No space left on device"
目录原因:cephfs的stray满了
stray就相当于回收站,stray有文件数限制,默认是10万,定时删除,但删除的很慢。导致stray一直处于满的状态。
ceph tell mds.node01 injectargs --mds_bal_fragment_size_max 1000000
然后把mds_bal_fragment_size_max写到配置文件里。
大功告成了。
https://cloud.tencent.com/developer/article/1177839
(没试过)
因为线上环境直接重启肯定不好。所以还是一边监测一边调整参数。
$ ceph daemonperf mds.node3 -----mds------ --mds_server-- ---objecter--- -----mds_cache----- ---mds_log---- rlat inos caps|hsr hcs hcr |writ read actv|recd recy stry purg|segs evts subm| 0 26k 129 | 0 0 0 | 0 0 0 | 0 0 6.2k 0 | 31 136 0 0 26k 129 | 0 0 0 | 0 0 0 | 0 0 6.2k 0 | 31 136 0
其中node3是自己active的mds的名称。
$ ceph daemon mds.node3 config show | grep mds_bal_fragment_size_max "mds_bal_fragment_size_max": "100000", #修改mds_bal_fragment_size_max $ ceph tell mds.node01 injectargs --mds_bal_fragment_size_max 1000000
查询命令还可以使用。"ceph --show-config"
同理修改其他配置:(有一点效果,但不推荐)
ceph tell mds.node03 injectargs --filer_max_purge_ops 10000 ceph tell mds.node03 injectargs --mds_max_purge_files 256
修改完后再次查看mds状态,会发现purge的数据频率会略微提升。
源码文件StrayManager.cc
/* * Note that there are compromises to how throttling * is implemented here, in the interests of simplicity: * * If insufficient ops are available to execute * the next item on the queue, we block even if * there are items further down the queue requiring * fewer ops which might be executable * * The ops considered "in use" by a purge will be * an overestimate over the period of execution, as * we count filer_max_purge_ops and ops for old backtraces * as in use throughout, even though towards the end * of the purge the actual ops in flight will be * lower. * * The ops limit may be exceeded if the number of ops * required by a single inode is greater than the * limit, for example directories with very many * fragments. */ /*......*/ // Work out a limit based on n_pgs / n_mdss, multiplied by the user's // preference for how many ops per PG max_purge_ops = uint64_t(((double)pg_count / (double)mds_count) * g_conf->mds_max_purge_ops_per_pg); // User may also specify a hard limit, apply this if so. if (g_conf->mds_max_purge_ops) { max_purge_ops = MIN(max_purge_ops, g_conf->mds_max_purge_ops); }
https://www.cnblogs.com/hlc-123/p/10700725.html
https://www.cnblogs.com/zphj1987/p/13575384.html
http://blog.wjin.org/posts/cephfs-remove-file-error.html
https://cloud.tencent.com/developer/article/1177839
https://docs.ceph.com/en/latest/cephfs/mds-config-ref/#
https://cloud.tencent.com/developer/article/1181883
https://tracker.ceph.com/issues/15920?journal_id=71023