Actions
Bug #14820
closeds3cmd rb --recursive doesn't decrease total_entries in stats
Status:
Duplicate
Priority:
Urgent
Assignee:
-
Target version:
-
% Done:
0%
Source:
Community (dev)
Tags:
Backport:
Regression:
No
Severity:
2 - major
Reviewed:
Affected Versions:
ceph-qa-suite:
Pull request ID:
Crash signature (v1):
Crash signature (v2):
Description
Hello,
When you use s3cmd with 'rb --recursive', the stats informations about user total_entries is not well decreased.
And if you have a 'max_bucket' limit in your quotas, you can anymore put files.
The problem seems come from the multiple delete.
How to reproduce:
- create an user
$> radosgw-admin user create --uid=johndoe --display-name="John Doe" $> radosgw-admin user stats --uid=johndoe --sync-stats ... "total_entries": 0, ...
- use s3cmd to sync "files" directory with your bucket
$> ls -l files | wc -l 30002 $> du -h files # 16k / file 469M files $> s3cmd mb s3://bucket $> s3cmd sync files s3://bucket ... ... ... ~3000 file -- perform ^C
- check user stats
$> radosgw-admin user stats --uid=johndoe --sync-stats ... "total_entries": 3000, ...
- delete the bucket
$> s3cmd rb --recursive s3://bucket
- check user stats
$> radosgw-admin user stats --uid=johndoe --sync-stats ... "total_entries": 45, # random value but not 0 as expected ...
- check my buckets
$> s3cmd ls s3:// # clean
- add max-objects limit
$> radosgw-admin quota set --quota-scope=user --uid=johndoe --max-objects=10 $> s3cmd mb s3://bucket $> s3cmd put files/file3 s3://bucket 'files/file3' -> 's3://bucket/file3' [1 of 1] 10240 of 10240 100% in 0s 2.64 MB/s done ERROR: S3 error: 403 (QuotaExceeded)
Tested with Hammer and Infernalis
ceph --version ceph version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299)
ceph --version ceph version 0.94.5 (9764da52395923e0b32908d83a9f7304401fee43)
Actions