Bug #4772
closed(deep?) scrubbing scheduling misses PGs
0%
Description
I have a 144 OSD (135 in) cluster, partioned in ~10 pools and 16760 pgs in total. The cluster runs Ceph 0.56.4 using Ubuntu 12.04.
As part of bug #4743, I had to run "for o in $osds; do ceph osd deep-scrub $o; done" on my cluster. I've set --osd-max-scrubs to 4, so about 140-150 pgs are getting scrubbed concurrently. This takes about 2 days, after which there are no pgs in deep scrubbing.
It turns out, though, that while most pgs did in fact get deep scrubbed, some of those were missed. In at least the two first rounds, possibly the third too. I'm afraid that grepping is hard, since I've basically run the three rounds too close to each other, so extracting more concrete numbers would be hard. The PGs being scrubbed are definitely in the right ballpark though, so I'm guessing we're talking about a few dozens PGs being missed.
This is obviously not very important but I'm filing it here as to not get completely forgotten :)