Bug #20280
openrgw: multi-site replication: switching master/secondary creates additional pool
0%
Description
We are testing multi-site replication between two clusters, plk041 and plk045, both on same ceph version 10.2.4 (9411351cc8ce9ee03fbd46225102fe3d28ddf611).
We have separate realm (replication), zonegroup (replication), zones (plk041-replication on plk041, plk045-replication on plk045), configured according to this doc: https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/2/html-single/object_gateway_guide_for_red_hat_enterprise_linux/#multi_site
To test switching master/secondary we are doing the following:
1. turning off master
2. setting secondary cluster as master
# radosgw-admin --cluster=plk045 --rgw-realm=replication zone modify --rgw-zone=plk045-replication --master # radosgw-admin --cluster=plk045 --rgw-realm=replication period update --commit
3. creating new bucket using new master (plk045 here)
After that we can see the new empty pool on plk045 cluster named plk041-replication.rgw.buckets.index,
ceph -c /etc/ceph/plk045.conf osd pool ls detail ... pool 129 'plk041-replication.rgw.buckets.index' replicated size 3 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 18861 flags hashpspool stripe_width 0
This issue is not stably reproduced, unfortunately. So we need some advice with debugging/reproducing it.
Files