// -*- mode:C++; tab-width:8; c-basic-offset:2; indent-tabs-mode:t -*- // vim: ts=8 sw=2 smarttab #include #include #include #include #include #include #include #include "auth/Crypto.h" #include "compressor/Compressor.h" #include "common/armor.h" #include "common/ceph_json.h" #include "common/config.h" #include "common/ceph_argparse.h" #include "common/Formatter.h" #include "common/errno.h" #include "common/safe_io.h" #include "cls/rgw/cls_rgw_client.h" #include "global/global_init.h" #include "include/utime.h" #include "include/str_list.h" #include "rgw_user.h" #include "rgw_bucket.h" #include "rgw_rados.h" #include "rgw_acl.h" #include "rgw_acl_s3.h" #include "rgw_lc.h" #include "rgw_log.h" #include "rgw_formats.h" #include "rgw_usage.h" #include "rgw_replica_log.h" #include "rgw_orphan.h" #include "rgw_sync.h" #include "rgw_sync_log_trim.h" #include "rgw_data_sync.h" #include "rgw_rest_conn.h" #include "rgw_realm_watcher.h" #include "rgw_role.h" #include "rgw_reshard.h" #include "rgw_http_client_curl.h" using namespace std; #define dout_context g_ceph_context #define dout_subsys ceph_subsys_rgw #define SECRET_KEY_LEN 40 #define PUBLIC_ID_LEN 20 static RGWRados *store = NULL; void usage() { cout << "usage: radosgw-admin [options...]" << std::endl; cout << "commands:\n"; cout << " user create create a new user\n" ; cout << " user modify modify user\n"; cout << " user info get user info\n"; cout << " user rm remove user\n"; cout << " user suspend suspend a user\n"; cout << " user enable re-enable user after suspension\n"; cout << " user check check user info\n"; cout << " user stats show user stats as accounted by quota subsystem\n"; cout << " user list list users\n"; cout << " caps add add user capabilities\n"; cout << " caps rm remove user capabilities\n"; cout << " subuser create create a new subuser\n" ; cout << " subuser modify modify subuser\n"; cout << " subuser rm remove subuser\n"; cout << " key create create access key\n"; cout << " key rm remove access key\n"; cout << " bucket list list buckets\n"; cout << " bucket limit check show bucket sharding stats\n"; cout << " bucket link link bucket to specified user\n"; cout << " bucket unlink unlink bucket from specified user\n"; cout << " bucket stats returns bucket statistics\n"; cout << " bucket rm remove bucket\n"; cout << " bucket check check bucket index\n"; cout << " bucket reshard reshard bucket\n"; cout << " bucket sync disable disable bucket sync\n"; cout << " bucket sync enable enable bucket sync\n"; cout << " bi get retrieve bucket index object entries\n"; cout << " bi put store bucket index object entries\n"; cout << " bi list list raw bucket index entries\n"; cout << " object rm remove object\n"; cout << " object stat stat an object for its metadata\n"; cout << " object unlink unlink object from bucket index\n"; cout << " objects expire run expired objects cleanup\n"; cout << " period delete delete a period\n"; cout << " period get get period info\n"; cout << " period get-current get current period info\n"; cout << " period pull pull a period\n"; cout << " period push push a period\n"; cout << " period list list all periods\n"; cout << " period update update the staging period\n"; cout << " period commit commit the staging period\n"; cout << " quota set set quota params\n"; cout << " quota enable enable quota\n"; cout << " quota disable disable quota\n"; cout << " global quota get view global quota params\n"; cout << " global quota set set global quota params\n"; cout << " global quota enable enable a global quota\n"; cout << " global quota disable disable a global quota\n"; cout << " realm create create a new realm\n"; cout << " realm delete delete a realm\n"; cout << " realm get show realm info\n"; cout << " realm get-default get default realm name\n"; cout << " realm list list realms\n"; cout << " realm list-periods list all realm periods\n"; cout << " realm remove remove a zonegroup from the realm\n"; cout << " realm rename rename a realm\n"; cout << " realm set set realm info (requires infile)\n"; cout << " realm default set realm as default\n"; cout << " realm pull pull a realm and its current period\n"; cout << " zonegroup add add a zone to a zonegroup\n"; cout << " zonegroup create create a new zone group info\n"; cout << " zonegroup default set default zone group\n"; cout << " zonegroup delete delete a zone group info\n"; cout << " zonegroup get show zone group info\n"; cout << " zonegroup modify modify an existing zonegroup\n"; cout << " zonegroup set set zone group info (requires infile)\n"; cout << " zonegroup remove remove a zone from a zonegroup\n"; cout << " zonegroup rename rename a zone group\n"; cout << " zonegroup list list all zone groups set on this cluster\n"; cout << " zonegroup placement list list zonegroup's placement targets\n"; cout << " zonegroup placement add add a placement target id to a zonegroup\n"; cout << " zonegroup placement modify modify a placement target of a specific zonegroup\n"; cout << " zonegroup placement rm remove a placement target from a zonegroup\n"; cout << " zonegroup placement default set a zonegroup's default placement target\n"; cout << " zone create create a new zone\n"; cout << " zone delete delete a zone\n"; cout << " zone get show zone cluster params\n"; cout << " zone modify modify an existing zone\n"; cout << " zone set set zone cluster params (requires infile)\n"; cout << " zone list list all zones set on this cluster\n"; cout << " zone rename rename a zone\n"; cout << " zone placement list list zone's placement targets\n"; cout << " zone placement add add a zone placement target\n"; cout << " zone placement modify modify a zone placement target\n"; cout << " zone placement rm remove a zone placement target\n"; cout << " pool add add an existing pool for data placement\n"; cout << " pool rm remove an existing pool from data placement set\n"; cout << " pools list list placement active set\n"; cout << " policy read bucket/object policy\n"; cout << " log list list log objects\n"; cout << " log show dump a log from specific object or (bucket + date\n"; cout << " + bucket-id)\n"; cout << " (NOTE: required to specify formatting of date\n"; cout << " to \"YYYY-MM-DD-hh\")\n"; cout << " log rm remove log object\n"; cout << " usage show show usage (by user, date range)\n"; cout << " usage trim trim usage (by user, date range)\n"; cout << " gc list dump expired garbage collection objects (specify\n"; cout << " --include-all to list all entries, including unexpired)\n"; cout << " gc process manually process garbage\n"; cout << " lc list list all bucket lifecycle progress\n"; cout << " lc process manually process lifecycle\n"; cout << " metadata get get metadata info\n"; cout << " metadata put put metadata info\n"; cout << " metadata rm remove metadata info\n"; cout << " metadata list list metadata info\n"; cout << " mdlog list list metadata log\n"; cout << " mdlog trim trim metadata log (use start-date, end-date or\n"; cout << " start-marker, end-marker)\n"; cout << " mdlog status read metadata log status\n"; cout << " bilog list list bucket index log\n"; cout << " bilog trim trim bucket index log (use start-marker, end-marker)\n"; cout << " datalog list list data log\n"; cout << " datalog trim trim data log\n"; cout << " datalog status read data log status\n"; cout << " opstate list list stateful operations entries (use client_id,\n"; cout << " op_id, object)\n"; cout << " opstate set set state on an entry (use client_id, op_id, object, state)\n"; cout << " opstate renew renew state on an entry (use client_id, op_id, object)\n"; cout << " opstate rm remove entry (use client_id, op_id, object)\n"; cout << " replicalog get get replica metadata log entry\n"; cout << " replicalog update update replica metadata log entry\n"; cout << " replicalog delete delete replica metadata log entry\n"; cout << " orphans find init and run search for leaked rados objects (use job-id, pool)\n"; cout << " orphans finish clean up search for leaked rados objects\n"; cout << " orphans list-jobs list the current job-ids for orphans search\n"; cout << " role create create a AWS role for use with STS\n"; cout << " role delete delete a role\n"; cout << " role get get a role\n"; cout << " role list list roles with specified path prefix\n"; cout << " role modify modify the assume role policy of an existing role\n"; cout << " role-policy put add/update permission policy to role\n"; cout << " role-policy list list policies attached to a role\n"; cout << " role-policy get get the specified inline policy document embedded with the given role\n"; cout << " role-policy delete delete policy attached to a role\n"; cout << " reshard add schedule a resharding of a bucket\n"; cout << " reshard list list all bucket resharding or scheduled to be reshared\n"; cout << " reshard process process of scheduled reshard jobs\n"; cout << " reshard cancel cancel resharding a bucket\n"; cout << " sync error list list sync error\n"; cout << " sync error trim trim sync error\n"; cout << "options:\n"; cout << " --tenant= tenant name\n"; cout << " --uid= user id\n"; cout << " --subuser= subuser name\n"; cout << " --access-key= S3 access key\n"; cout << " --email=\n"; cout << " --secret/--secret-key=\n"; cout << " specify secret key\n"; cout << " --gen-access-key generate random access key (for S3)\n"; cout << " --gen-secret generate random secret key\n"; cout << " --key-type= key type, options are: swift, s3\n"; cout << " --temp-url-key[-2]= temp url key\n"; cout << " --access= Set access permissions for sub-user, should be one\n"; cout << " of read, write, readwrite, full\n"; cout << " --display-name=\n"; cout << " --max-buckets max number of buckets for a user\n"; cout << " --admin set the admin flag on the user\n"; cout << " --system set the system flag on the user\n"; cout << " --bucket= Specify the bucket name. Also used by the quota command.\n"; cout << " --pool= Specify the pool name. Also used to scan for leaked rados objects.\n"; cout << " --object= object name\n"; cout << " --date= date in the format yyyy-mm-dd\n"; cout << " --start-date= start date in the format yyyy-mm-dd\n"; cout << " --end-date= end date in the format yyyy-mm-dd\n"; cout << " --bucket-id= bucket id\n"; cout << " --shard-id= optional for: \n"; cout << " mdlog list\n"; cout << " data sync status\n"; cout << " required for: \n"; cout << " mdlog trim\n"; cout << " replica mdlog get/delete\n"; cout << " replica datalog get/delete\n"; cout << " --max-entries= max entries for listing operations\n"; cout << " --metadata-key= key to retrieve metadata from with metadata get\n"; cout << " --remote= zone or zonegroup id of remote gateway\n"; cout << " --period= period id\n"; cout << " --epoch= period epoch\n"; cout << " --commit commit the period during 'period update'\n"; cout << " --staging get staging period info\n"; cout << " --master set as master\n"; cout << " --master-url master url\n"; cout << " --master-zonegroup= master zonegroup id\n"; cout << " --master-zone= master zone id\n"; cout << " --rgw-realm= realm name\n"; cout << " --realm-id= realm id\n"; cout << " --realm-new-name= realm new name\n"; cout << " --rgw-zonegroup= zonegroup name\n"; cout << " --zonegroup-id= zonegroup id\n"; cout << " --zonegroup-new-name=\n"; cout << " zonegroup new name\n"; cout << " --rgw-zone= name of zone in which radosgw is running\n"; cout << " --zone-id= zone id\n"; cout << " --zone-new-name= zone new name\n"; cout << " --source-zone specify the source zone (for data sync)\n"; cout << " --default set entity (realm, zonegroup, zone) as default\n"; cout << " --read-only set zone as read-only (when adding to zonegroup)\n"; cout << " --placement-id placement id for zonegroup placement commands\n"; cout << " --tags= list of tags for zonegroup placement add and modify commands\n"; cout << " --tags-add= list of tags to add for zonegroup placement modify command\n"; cout << " --tags-rm= list of tags to remove for zonegroup placement modify command\n"; cout << " --endpoints= zone endpoints\n"; cout << " --index-pool= placement target index pool\n"; cout << " --data-pool= placement target data pool\n"; cout << " --data-extra-pool= placement target data extra (non-ec) pool\n"; cout << " --placement-index-type=\n"; cout << " placement target index type (normal, indexless, or #id)\n"; cout << " --compression= placement target compression type (plugin name or empty/none)\n"; cout << " --tier-type= zone tier type\n"; cout << " --tier-config==[,...]\n"; cout << " set zone tier config keys, values\n"; cout << " --tier-config-rm=[,...]\n"; cout << " unset zone tier config keys\n"; cout << " --sync-from-all[=false] set/reset whether zone syncs from all zonegroup peers\n"; cout << " --sync-from=[zone-name][,...]\n"; cout << " set list of zones to sync from\n"; cout << " --sync-from-rm=[zone-name][,...]\n"; cout << " remove zones from list of zones to sync from\n"; cout << " --fix besides checking bucket index, will also fix it\n"; cout << " --check-objects bucket check: rebuilds bucket index according to\n"; cout << " actual objects state\n"; cout << " --format= specify output format for certain operations: xml,\n"; cout << " json\n"; cout << " --purge-data when specified, user removal will also purge all the\n"; cout << " user data\n"; cout << " --purge-keys when specified, subuser removal will also purge all the\n"; cout << " subuser keys\n"; cout << " --purge-objects remove a bucket's objects before deleting it\n"; cout << " (NOTE: required to delete a non-empty bucket)\n"; cout << " --sync-stats option to 'user stats', update user stats with current\n"; cout << " stats reported by user's buckets indexes\n"; cout << " --reset-stats option to 'user stats', reset stats in accordance with user buckets\n"; cout << " --show-log-entries= enable/disable dump of log entries on log show\n"; cout << " --show-log-sum= enable/disable dump of log summation on log show\n"; cout << " --skip-zero-entries log show only dumps entries that don't have zero value\n"; cout << " in one of the numeric field\n"; cout << " --infile= specify a file to read in when setting data\n"; cout << " --state= specify a state for the opstate set command\n"; cout << " --replica-log-type replica log type (metadata, data, bucket), required for\n"; cout << " replica log operations\n"; cout << " --categories= comma separated list of categories, used in usage show\n"; cout << " --caps= list of caps (e.g., \"usage=read, write; user=read\")\n"; cout << " --yes-i-really-mean-it required for certain operations\n"; cout << " --warnings-only when specified with bucket limit check, list\n"; cout << " only buckets nearing or over the current max\n"; cout << " objects per shard value\n"; cout << " --bypass-gc when specified with bucket deletion, triggers\n"; cout << " object deletions by not involving GC\n"; cout << " --inconsistent-index when specified with bucket deletion and bypass-gc set to true,\n"; cout << " ignores bucket index consistency\n"; cout << "\n"; cout << " := \"YYYY-MM-DD[ hh:mm:ss]\"\n"; cout << "\nQuota options:\n"; cout << " --bucket specified bucket for quota command\n"; cout << " --max-objects specify max objects (negative value to disable)\n"; cout << " --max-size specify max size (in B/K/M/G/T, negative value to disable)\n"; cout << " --quota-scope scope of quota (bucket, user)\n"; cout << "\nOrphans search options:\n"; cout << " --pool data pool to scan for leaked rados objects in\n"; cout << " --num-shards num of shards to use for keeping the temporary scan info\n"; cout << " --orphan-stale-secs num of seconds to wait before declaring an object to be an orphan (default: 86400)\n"; cout << " --job-id set the job id (for orphans find)\n"; cout << " --max-concurrent-ios maximum concurrent ios for orphans find (default: 32)\n"; cout << "\nOrphans list-jobs options:\n"; cout << " --extra-info provide extra info in job list\n"; cout << "\nRole options:\n"; cout << " --role-name name of the role to create\n"; cout << " --path path to the role\n"; cout << " --assume-role-policy-doc the trust relationship policy document that grants an entity permission to assume the role\n"; cout << " --policy-name name of the policy document\n"; cout << " --policy-doc permission policy document\n"; cout << " --path-prefix path prefix for filtering roles\n"; cout << "\n"; generic_client_usage(); } enum { OPT_NO_CMD = 0, OPT_USER_CREATE, OPT_USER_INFO, OPT_USER_MODIFY, OPT_USER_RM, OPT_USER_SUSPEND, OPT_USER_ENABLE, OPT_USER_CHECK, OPT_USER_STATS, OPT_USER_LIST, OPT_SUBUSER_CREATE, OPT_SUBUSER_MODIFY, OPT_SUBUSER_RM, OPT_KEY_CREATE, OPT_KEY_RM, OPT_BUCKETS_LIST, OPT_BUCKET_LIMIT_CHECK, OPT_BUCKET_LINK, OPT_BUCKET_UNLINK, OPT_BUCKET_STATS, OPT_BUCKET_CHECK, OPT_BUCKET_SYNC_STATUS, OPT_BUCKET_SYNC_MARKERS, OPT_BUCKET_SYNC_INIT, OPT_BUCKET_SYNC_RUN, OPT_BUCKET_SYNC_DISABLE, OPT_BUCKET_SYNC_ENABLE, OPT_BUCKET_RM, OPT_BUCKET_REWRITE, OPT_BUCKET_RESHARD, OPT_POLICY, OPT_POOL_ADD, OPT_POOL_RM, OPT_POOLS_LIST, OPT_LOG_LIST, OPT_LOG_SHOW, OPT_LOG_RM, OPT_USAGE_SHOW, OPT_USAGE_TRIM, OPT_OBJECT_RM, OPT_OBJECT_UNLINK, OPT_OBJECT_STAT, OPT_OBJECT_REWRITE, OPT_OBJECTS_EXPIRE, OPT_BI_GET, OPT_BI_PUT, OPT_BI_LIST, OPT_BI_PURGE, OPT_OLH_GET, OPT_OLH_READLOG, OPT_QUOTA_SET, OPT_QUOTA_ENABLE, OPT_QUOTA_DISABLE, OPT_GC_LIST, OPT_GC_PROCESS, OPT_LC_LIST, OPT_LC_PROCESS, OPT_ORPHANS_FIND, OPT_ORPHANS_FINISH, OPT_ORPHANS_LIST_JOBS, OPT_ZONEGROUP_ADD, OPT_ZONEGROUP_CREATE, OPT_ZONEGROUP_DEFAULT, OPT_ZONEGROUP_DELETE, OPT_ZONEGROUP_GET, OPT_ZONEGROUP_MODIFY, OPT_ZONEGROUP_SET, OPT_ZONEGROUP_LIST, OPT_ZONEGROUP_REMOVE, OPT_ZONEGROUP_RENAME, OPT_ZONEGROUP_PLACEMENT_ADD, OPT_ZONEGROUP_PLACEMENT_MODIFY, OPT_ZONEGROUP_PLACEMENT_RM, OPT_ZONEGROUP_PLACEMENT_LIST, OPT_ZONEGROUP_PLACEMENT_DEFAULT, OPT_ZONE_CREATE, OPT_ZONE_DELETE, OPT_ZONE_GET, OPT_ZONE_MODIFY, OPT_ZONE_SET, OPT_ZONE_LIST, OPT_ZONE_RENAME, OPT_ZONE_DEFAULT, OPT_ZONE_PLACEMENT_ADD, OPT_ZONE_PLACEMENT_MODIFY, OPT_ZONE_PLACEMENT_RM, OPT_ZONE_PLACEMENT_LIST, OPT_CAPS_ADD, OPT_CAPS_RM, OPT_METADATA_GET, OPT_METADATA_PUT, OPT_METADATA_RM, OPT_METADATA_LIST, OPT_METADATA_SYNC_STATUS, OPT_METADATA_SYNC_INIT, OPT_METADATA_SYNC_RUN, OPT_MDLOG_LIST, OPT_MDLOG_AUTOTRIM, OPT_MDLOG_TRIM, OPT_MDLOG_FETCH, OPT_MDLOG_STATUS, OPT_SYNC_ERROR_LIST, OPT_SYNC_ERROR_TRIM, OPT_BILOG_LIST, OPT_BILOG_TRIM, OPT_BILOG_STATUS, OPT_BILOG_AUTOTRIM, OPT_DATA_SYNC_STATUS, OPT_DATA_SYNC_INIT, OPT_DATA_SYNC_RUN, OPT_DATALOG_LIST, OPT_DATALOG_STATUS, OPT_DATALOG_TRIM, OPT_OPSTATE_LIST, OPT_OPSTATE_SET, OPT_OPSTATE_RENEW, OPT_OPSTATE_RM, OPT_REPLICALOG_GET, OPT_REPLICALOG_UPDATE, OPT_REPLICALOG_DELETE, OPT_REALM_CREATE, OPT_REALM_DELETE, OPT_REALM_GET, OPT_REALM_GET_DEFAULT, OPT_REALM_LIST, OPT_REALM_LIST_PERIODS, OPT_REALM_REMOVE, OPT_REALM_RENAME, OPT_REALM_SET, OPT_REALM_DEFAULT, OPT_REALM_PULL, OPT_PERIOD_DELETE, OPT_PERIOD_GET, OPT_PERIOD_GET_CURRENT, OPT_PERIOD_PULL, OPT_PERIOD_PUSH, OPT_PERIOD_LIST, OPT_PERIOD_UPDATE, OPT_PERIOD_COMMIT, OPT_GLOBAL_QUOTA_GET, OPT_GLOBAL_QUOTA_SET, OPT_GLOBAL_QUOTA_ENABLE, OPT_GLOBAL_QUOTA_DISABLE, OPT_SYNC_STATUS, OPT_ROLE_CREATE, OPT_ROLE_DELETE, OPT_ROLE_GET, OPT_ROLE_MODIFY, OPT_ROLE_LIST, OPT_ROLE_POLICY_PUT, OPT_ROLE_POLICY_LIST, OPT_ROLE_POLICY_GET, OPT_ROLE_POLICY_DELETE, OPT_RESHARD_ADD, OPT_RESHARD_LIST, OPT_RESHARD_STATUS, OPT_RESHARD_PROCESS, OPT_RESHARD_CANCEL, }; static int get_cmd(const char *cmd, const char *prev_cmd, const char *prev_prev_cmd, bool *need_more) { *need_more = false; // NOTE: please keep the checks in alphabetical order !!! if (strcmp(cmd, "bi") == 0 || strcmp(cmd, "bilog") == 0 || strcmp(cmd, "buckets") == 0 || strcmp(cmd, "caps") == 0 || strcmp(cmd, "data") == 0 || strcmp(cmd, "datalog") == 0 || strcmp(cmd, "error") == 0 || strcmp(cmd, "gc") == 0 || strcmp(cmd, "global") == 0 || strcmp(cmd, "key") == 0 || strcmp(cmd, "log") == 0 || strcmp(cmd, "lc") == 0 || strcmp(cmd, "mdlog") == 0 || strcmp(cmd, "metadata") == 0 || strcmp(cmd, "object") == 0 || strcmp(cmd, "objects") == 0 || strcmp(cmd, "olh") == 0 || strcmp(cmd, "opstate") == 0 || strcmp(cmd, "orphans") == 0 || strcmp(cmd, "period") == 0 || strcmp(cmd, "placement") == 0 || strcmp(cmd, "pool") == 0 || strcmp(cmd, "pools") == 0 || strcmp(cmd, "quota") == 0 || strcmp(cmd, "realm") == 0 || strcmp(cmd, "replicalog") == 0 || strcmp(cmd, "role") == 0 || strcmp(cmd, "role-policy") == 0 || strcmp(cmd, "subuser") == 0 || strcmp(cmd, "sync") == 0 || strcmp(cmd, "usage") == 0 || strcmp(cmd, "user") == 0 || strcmp(cmd, "zone") == 0 || strcmp(cmd, "zonegroup") == 0 || strcmp(cmd, "zonegroups") == 0) { *need_more = true; return 0; } /* * can do both radosgw-admin bucket reshard, and radosgw-admin reshard bucket */ if (strcmp(cmd, "reshard") == 0 && !(prev_cmd && strcmp(prev_cmd, "bucket") == 0)) { *need_more = true; return 0; } if (strcmp(cmd, "bucket") == 0 && !(prev_cmd && strcmp(prev_cmd, "reshard") == 0)) { *need_more = true; return 0; } if (strcmp(cmd, "policy") == 0) return OPT_POLICY; if (!prev_cmd) return -EINVAL; if (strcmp(prev_cmd, "user") == 0) { if (strcmp(cmd, "create") == 0) return OPT_USER_CREATE; if (strcmp(cmd, "info") == 0) return OPT_USER_INFO; if (strcmp(cmd, "modify") == 0) return OPT_USER_MODIFY; if (strcmp(cmd, "rm") == 0) return OPT_USER_RM; if (strcmp(cmd, "suspend") == 0) return OPT_USER_SUSPEND; if (strcmp(cmd, "enable") == 0) return OPT_USER_ENABLE; if (strcmp(cmd, "check") == 0) return OPT_USER_CHECK; if (strcmp(cmd, "stats") == 0) return OPT_USER_STATS; if (strcmp(cmd, "list") == 0) return OPT_USER_LIST; } else if (strcmp(prev_cmd, "subuser") == 0) { if (strcmp(cmd, "create") == 0) return OPT_SUBUSER_CREATE; if (strcmp(cmd, "modify") == 0) return OPT_SUBUSER_MODIFY; if (strcmp(cmd, "rm") == 0) return OPT_SUBUSER_RM; } else if (strcmp(prev_cmd, "key") == 0) { if (strcmp(cmd, "create") == 0) return OPT_KEY_CREATE; if (strcmp(cmd, "rm") == 0) return OPT_KEY_RM; } else if (strcmp(prev_cmd, "buckets") == 0) { if (strcmp(cmd, "list") == 0) return OPT_BUCKETS_LIST; } else if (strcmp(prev_cmd, "bucket") == 0) { if (strcmp(cmd, "list") == 0) return OPT_BUCKETS_LIST; if (strcmp(cmd, "link") == 0) return OPT_BUCKET_LINK; if (strcmp(cmd, "unlink") == 0) return OPT_BUCKET_UNLINK; if (strcmp(cmd, "stats") == 0) return OPT_BUCKET_STATS; if (strcmp(cmd, "rm") == 0) return OPT_BUCKET_RM; if (strcmp(cmd, "rewrite") == 0) return OPT_BUCKET_REWRITE; if (strcmp(cmd, "reshard") == 0) return OPT_BUCKET_RESHARD; if (strcmp(cmd, "check") == 0) return OPT_BUCKET_CHECK; if (strcmp(cmd, "sync") == 0) { *need_more = true; return 0; } if (strcmp(cmd, "limit") == 0) { *need_more = true; return 0; } } else if (prev_prev_cmd && strcmp(prev_prev_cmd, "bucket") == 0) { if (strcmp(prev_cmd, "sync") == 0) { if (strcmp(cmd, "status") == 0) return OPT_BUCKET_SYNC_STATUS; if (strcmp(cmd, "markers") == 0) return OPT_BUCKET_SYNC_MARKERS; if (strcmp(cmd, "init") == 0) return OPT_BUCKET_SYNC_INIT; if (strcmp(cmd, "run") == 0) return OPT_BUCKET_SYNC_RUN; if (strcmp(cmd, "disable") == 0) return OPT_BUCKET_SYNC_DISABLE; if (strcmp(cmd, "enable") == 0) return OPT_BUCKET_SYNC_ENABLE; } else if ((strcmp(prev_cmd, "limit") == 0) && (strcmp(cmd, "check") == 0)) { return OPT_BUCKET_LIMIT_CHECK; } } else if (strcmp(prev_cmd, "log") == 0) { if (strcmp(cmd, "list") == 0) return OPT_LOG_LIST; if (strcmp(cmd, "show") == 0) return OPT_LOG_SHOW; if (strcmp(cmd, "rm") == 0) return OPT_LOG_RM; } else if (strcmp(prev_cmd, "usage") == 0) { if (strcmp(cmd, "show") == 0) return OPT_USAGE_SHOW; if (strcmp(cmd, "trim") == 0) return OPT_USAGE_TRIM; } else if (strcmp(prev_cmd, "caps") == 0) { if (strcmp(cmd, "add") == 0) return OPT_CAPS_ADD; if (strcmp(cmd, "rm") == 0) return OPT_CAPS_RM; } else if (strcmp(prev_cmd, "pool") == 0) { if (strcmp(cmd, "add") == 0) return OPT_POOL_ADD; if (strcmp(cmd, "rm") == 0) return OPT_POOL_RM; if (strcmp(cmd, "list") == 0) return OPT_POOLS_LIST; } else if (strcmp(prev_cmd, "pools") == 0) { if (strcmp(cmd, "list") == 0) return OPT_POOLS_LIST; } else if (strcmp(prev_cmd, "object") == 0) { if (strcmp(cmd, "rm") == 0) return OPT_OBJECT_RM; if (strcmp(cmd, "unlink") == 0) return OPT_OBJECT_UNLINK; if (strcmp(cmd, "stat") == 0) return OPT_OBJECT_STAT; if (strcmp(cmd, "rewrite") == 0) return OPT_OBJECT_REWRITE; } else if (strcmp(prev_cmd, "objects") == 0) { if (strcmp(cmd, "expire") == 0) return OPT_OBJECTS_EXPIRE; } else if (strcmp(prev_cmd, "olh") == 0) { if (strcmp(cmd, "get") == 0) return OPT_OLH_GET; if (strcmp(cmd, "readlog") == 0) return OPT_OLH_READLOG; } else if (strcmp(prev_cmd, "bi") == 0) { if (strcmp(cmd, "get") == 0) return OPT_BI_GET; if (strcmp(cmd, "put") == 0) return OPT_BI_PUT; if (strcmp(cmd, "list") == 0) return OPT_BI_LIST; if (strcmp(cmd, "purge") == 0) return OPT_BI_PURGE; } else if ((prev_prev_cmd && strcmp(prev_prev_cmd, "global") == 0) && (strcmp(prev_cmd, "quota") == 0)) { if (strcmp(cmd, "get") == 0) return OPT_GLOBAL_QUOTA_GET; if (strcmp(cmd, "set") == 0) return OPT_GLOBAL_QUOTA_SET; if (strcmp(cmd, "enable") == 0) return OPT_GLOBAL_QUOTA_ENABLE; if (strcmp(cmd, "disable") == 0) return OPT_GLOBAL_QUOTA_DISABLE; } else if (strcmp(prev_cmd, "period") == 0) { if (strcmp(cmd, "delete") == 0) return OPT_PERIOD_DELETE; if (strcmp(cmd, "get") == 0) return OPT_PERIOD_GET; if (strcmp(cmd, "get-current") == 0) return OPT_PERIOD_GET_CURRENT; if (strcmp(cmd, "pull") == 0) return OPT_PERIOD_PULL; if (strcmp(cmd, "push") == 0) return OPT_PERIOD_PUSH; if (strcmp(cmd, "list") == 0) return OPT_PERIOD_LIST; if (strcmp(cmd, "update") == 0) return OPT_PERIOD_UPDATE; if (strcmp(cmd, "commit") == 0) return OPT_PERIOD_COMMIT; } else if (strcmp(prev_cmd, "realm") == 0) { if (strcmp(cmd, "create") == 0) return OPT_REALM_CREATE; if (strcmp(cmd, "delete") == 0) return OPT_REALM_DELETE; if (strcmp(cmd, "get") == 0) return OPT_REALM_GET; if (strcmp(cmd, "get-default") == 0) return OPT_REALM_GET_DEFAULT; if (strcmp(cmd, "list") == 0) return OPT_REALM_LIST; if (strcmp(cmd, "list-periods") == 0) return OPT_REALM_LIST_PERIODS; if (strcmp(cmd, "remove") == 0) return OPT_REALM_REMOVE; if (strcmp(cmd, "rename") == 0) return OPT_REALM_RENAME; if (strcmp(cmd, "set") == 0) return OPT_REALM_SET; if (strcmp(cmd, "default") == 0) return OPT_REALM_DEFAULT; if (strcmp(cmd, "pull") == 0) return OPT_REALM_PULL; } else if ((prev_prev_cmd && strcmp(prev_prev_cmd, "zonegroup") == 0) && (strcmp(prev_cmd, "placement") == 0)) { if (strcmp(cmd, "add") == 0) return OPT_ZONEGROUP_PLACEMENT_ADD; if (strcmp(cmd, "modify") == 0) return OPT_ZONEGROUP_PLACEMENT_MODIFY; if (strcmp(cmd, "rm") == 0) return OPT_ZONEGROUP_PLACEMENT_RM; if (strcmp(cmd, "list") == 0) return OPT_ZONEGROUP_PLACEMENT_LIST; if (strcmp(cmd, "default") == 0) return OPT_ZONEGROUP_PLACEMENT_DEFAULT; } else if (strcmp(prev_cmd, "zonegroup") == 0) { if (strcmp(cmd, "add") == 0) return OPT_ZONEGROUP_ADD; if (strcmp(cmd, "create")== 0) return OPT_ZONEGROUP_CREATE; if (strcmp(cmd, "default") == 0) return OPT_ZONEGROUP_DEFAULT; if (strcmp(cmd, "delete") == 0) return OPT_ZONEGROUP_DELETE; if (strcmp(cmd, "get") == 0) return OPT_ZONEGROUP_GET; if (strcmp(cmd, "modify") == 0) return OPT_ZONEGROUP_MODIFY; if (strcmp(cmd, "list") == 0) return OPT_ZONEGROUP_LIST; if (strcmp(cmd, "set") == 0) return OPT_ZONEGROUP_SET; if (strcmp(cmd, "remove") == 0) return OPT_ZONEGROUP_REMOVE; if (strcmp(cmd, "rename") == 0) return OPT_ZONEGROUP_RENAME; } else if (strcmp(prev_cmd, "quota") == 0) { if (strcmp(cmd, "set") == 0) return OPT_QUOTA_SET; if (strcmp(cmd, "enable") == 0) return OPT_QUOTA_ENABLE; if (strcmp(cmd, "disable") == 0) return OPT_QUOTA_DISABLE; } else if (strcmp(prev_cmd, "zonegroups") == 0) { if (strcmp(cmd, "list") == 0) return OPT_ZONEGROUP_LIST; } else if ((prev_prev_cmd && strcmp(prev_prev_cmd, "zone") == 0) && (strcmp(prev_cmd, "placement") == 0)) { if (strcmp(cmd, "add") == 0) return OPT_ZONE_PLACEMENT_ADD; if (strcmp(cmd, "modify") == 0) return OPT_ZONE_PLACEMENT_MODIFY; if (strcmp(cmd, "rm") == 0) return OPT_ZONE_PLACEMENT_RM; if (strcmp(cmd, "list") == 0) return OPT_ZONE_PLACEMENT_LIST; } else if (strcmp(prev_cmd, "zone") == 0) { if (strcmp(cmd, "delete") == 0) return OPT_ZONE_DELETE; if (strcmp(cmd, "create") == 0) return OPT_ZONE_CREATE; if (strcmp(cmd, "get") == 0) return OPT_ZONE_GET; if (strcmp(cmd, "set") == 0) return OPT_ZONE_SET; if (strcmp(cmd, "list") == 0) return OPT_ZONE_LIST; if (strcmp(cmd, "modify") == 0) return OPT_ZONE_MODIFY; if (strcmp(cmd, "rename") == 0) return OPT_ZONE_RENAME; if (strcmp(cmd, "default") == 0) return OPT_ZONE_DEFAULT; } else if (strcmp(prev_cmd, "zones") == 0) { if (strcmp(cmd, "list") == 0) return OPT_ZONE_LIST; } else if (strcmp(prev_cmd, "gc") == 0) { if (strcmp(cmd, "list") == 0) return OPT_GC_LIST; if (strcmp(cmd, "process") == 0) return OPT_GC_PROCESS; } else if (strcmp(prev_cmd, "lc") == 0) { if (strcmp(cmd, "list") == 0) return OPT_LC_LIST; if (strcmp(cmd, "process") == 0) return OPT_LC_PROCESS; } else if (strcmp(prev_cmd, "orphans") == 0) { if (strcmp(cmd, "find") == 0) return OPT_ORPHANS_FIND; if (strcmp(cmd, "finish") == 0) return OPT_ORPHANS_FINISH; if (strcmp(cmd, "list-jobs") == 0) return OPT_ORPHANS_LIST_JOBS; } else if (strcmp(prev_cmd, "metadata") == 0) { if (strcmp(cmd, "get") == 0) return OPT_METADATA_GET; if (strcmp(cmd, "put") == 0) return OPT_METADATA_PUT; if (strcmp(cmd, "rm") == 0) return OPT_METADATA_RM; if (strcmp(cmd, "list") == 0) return OPT_METADATA_LIST; if (strcmp(cmd, "sync") == 0) { *need_more = true; return 0; } } else if ((prev_prev_cmd && strcmp(prev_prev_cmd, "metadata") == 0) && (strcmp(prev_cmd, "sync") == 0)) { if (strcmp(cmd, "status") == 0) return OPT_METADATA_SYNC_STATUS; if (strcmp(cmd, "init") == 0) return OPT_METADATA_SYNC_INIT; if (strcmp(cmd, "run") == 0) return OPT_METADATA_SYNC_RUN; } else if ((prev_prev_cmd && strcmp(prev_prev_cmd, "sync") == 0) && (strcmp(prev_cmd, "error") == 0)) { if (strcmp(cmd, "list") == 0) return OPT_SYNC_ERROR_LIST; if (strcmp(cmd, "trim") == 0) return OPT_SYNC_ERROR_TRIM; } else if (strcmp(prev_cmd, "mdlog") == 0) { if (strcmp(cmd, "list") == 0) return OPT_MDLOG_LIST; if (strcmp(cmd, "autotrim") == 0) return OPT_MDLOG_AUTOTRIM; if (strcmp(cmd, "trim") == 0) return OPT_MDLOG_TRIM; if (strcmp(cmd, "fetch") == 0) return OPT_MDLOG_FETCH; if (strcmp(cmd, "status") == 0) return OPT_MDLOG_STATUS; } else if (strcmp(prev_cmd, "bilog") == 0) { if (strcmp(cmd, "list") == 0) return OPT_BILOG_LIST; if (strcmp(cmd, "trim") == 0) return OPT_BILOG_TRIM; if (strcmp(cmd, "status") == 0) return OPT_BILOG_STATUS; if (strcmp(cmd, "autotrim") == 0) return OPT_BILOG_AUTOTRIM; } else if (strcmp(prev_cmd, "data") == 0) { if (strcmp(cmd, "sync") == 0) { *need_more = true; return 0; } } else if (strcmp(prev_cmd, "datalog") == 0) { if (strcmp(cmd, "list") == 0) return OPT_DATALOG_LIST; if (strcmp(cmd, "trim") == 0) return OPT_DATALOG_TRIM; if (strcmp(cmd, "status") == 0) return OPT_DATALOG_STATUS; } else if ((prev_prev_cmd && strcmp(prev_prev_cmd, "data") == 0) && (strcmp(prev_cmd, "sync") == 0)) { if (strcmp(cmd, "status") == 0) return OPT_DATA_SYNC_STATUS; if (strcmp(cmd, "init") == 0) return OPT_DATA_SYNC_INIT; if (strcmp(cmd, "run") == 0) return OPT_DATA_SYNC_RUN; } else if (strcmp(prev_cmd, "opstate") == 0) { if (strcmp(cmd, "list") == 0) return OPT_OPSTATE_LIST; if (strcmp(cmd, "set") == 0) return OPT_OPSTATE_SET; if (strcmp(cmd, "renew") == 0) return OPT_OPSTATE_RENEW; if (strcmp(cmd, "rm") == 0) return OPT_OPSTATE_RM; } else if (strcmp(prev_cmd, "replicalog") == 0) { if (strcmp(cmd, "get") == 0) return OPT_REPLICALOG_GET; if (strcmp(cmd, "update") == 0) return OPT_REPLICALOG_UPDATE; if (strcmp(cmd, "delete") == 0) return OPT_REPLICALOG_DELETE; } else if (strcmp(prev_cmd, "sync") == 0) { if (strcmp(cmd, "status") == 0) return OPT_SYNC_STATUS; } else if (strcmp(prev_cmd, "role") == 0) { if (strcmp(cmd, "create") == 0) return OPT_ROLE_CREATE; if (strcmp(cmd, "delete") == 0) return OPT_ROLE_DELETE; if (strcmp(cmd, "get") == 0) return OPT_ROLE_GET; if (strcmp(cmd, "modify") == 0) return OPT_ROLE_MODIFY; if (strcmp(cmd, "list") == 0) return OPT_ROLE_LIST; } else if (strcmp(prev_cmd, "role-policy") == 0) { if (strcmp(cmd, "put") == 0) return OPT_ROLE_POLICY_PUT; if (strcmp(cmd, "list") == 0) return OPT_ROLE_POLICY_LIST; if (strcmp(cmd, "get") == 0) return OPT_ROLE_POLICY_GET; if (strcmp(cmd, "delete") == 0) return OPT_ROLE_POLICY_DELETE; } else if (strcmp(prev_cmd, "reshard") == 0) { if (strcmp(cmd, "bucket") == 0) return OPT_BUCKET_RESHARD; if (strcmp(cmd, "add") == 0) return OPT_RESHARD_ADD; if (strcmp(cmd, "list") == 0) return OPT_RESHARD_LIST; if (strcmp(cmd, "status") == 0) return OPT_RESHARD_STATUS; if (strcmp(cmd, "process") == 0) return OPT_RESHARD_PROCESS; if (strcmp(cmd, "cancel") == 0) return OPT_RESHARD_CANCEL; } return -EINVAL; } enum ReplicaLogType { ReplicaLog_Invalid = 0, ReplicaLog_Metadata, ReplicaLog_Data, ReplicaLog_Bucket, }; ReplicaLogType get_replicalog_type(const string& name) { if (name == "md" || name == "meta" || name == "metadata") return ReplicaLog_Metadata; if (name == "data") return ReplicaLog_Data; if (name == "bucket") return ReplicaLog_Bucket; return ReplicaLog_Invalid; } BIIndexType get_bi_index_type(const string& type_str) { if (type_str == "plain") return PlainIdx; if (type_str == "instance") return InstanceIdx; if (type_str == "olh") return OLHIdx; return InvalidIdx; } void dump_bi_entry(bufferlist& bl, BIIndexType index_type, Formatter *formatter) { bufferlist::iterator iter = bl.begin(); switch (index_type) { case PlainIdx: case InstanceIdx: { rgw_bucket_dir_entry entry; ::decode(entry, iter); encode_json("entry", entry, formatter); } break; case OLHIdx: { rgw_bucket_olh_entry entry; ::decode(entry, iter); encode_json("entry", entry, formatter); } break; default: ceph_abort(); break; } } static void show_user_info(RGWUserInfo& info, Formatter *formatter) { encode_json("user_info", info, formatter); formatter->flush(cout); cout << std::endl; } static void show_perm_policy(string perm_policy, Formatter* formatter) { formatter->open_object_section("role"); formatter->dump_string("Permission policy", perm_policy); formatter->close_section(); formatter->flush(cout); } static void show_policy_names(std::vector policy_names, Formatter* formatter) { formatter->open_array_section("PolicyNames"); for (const auto& it : policy_names) { formatter->dump_string("policyname", it); } formatter->close_section(); formatter->flush(cout); } static void show_role_info(RGWRole& role, Formatter* formatter) { formatter->open_object_section("role"); role.dump(formatter); formatter->close_section(); formatter->flush(cout); } static void show_roles_info(vector& roles, Formatter* formatter) { formatter->open_array_section("Roles"); for (const auto& it : roles) { formatter->open_object_section("role"); it.dump(formatter); formatter->close_section(); } formatter->close_section(); formatter->flush(cout); } class StoreDestructor { RGWRados *store; public: explicit StoreDestructor(RGWRados *_s) : store(_s) {} ~StoreDestructor() { RGWStoreManager::close_storage(store); } }; static int init_bucket(const string& tenant_name, const string& bucket_name, const string& bucket_id, RGWBucketInfo& bucket_info, rgw_bucket& bucket, map *pattrs = nullptr) { if (!bucket_name.empty()) { RGWObjectCtx obj_ctx(store); int r; if (bucket_id.empty()) { r = store->get_bucket_info(obj_ctx, tenant_name, bucket_name, bucket_info, nullptr, pattrs); } else { string bucket_instance_id = bucket_name + ":" + bucket_id; r = store->get_bucket_instance_info(obj_ctx, bucket_instance_id, bucket_info, NULL, pattrs); } if (r < 0) { cerr << "could not get bucket info for bucket=" << bucket_name << std::endl; return r; } bucket = bucket_info.bucket; } return 0; } static int read_input(const string& infile, bufferlist& bl) { int fd = 0; if (infile.size()) { fd = open(infile.c_str(), O_RDONLY); if (fd < 0) { int err = -errno; cerr << "error reading input file " << infile << std::endl; return err; } } #define READ_CHUNK 8196 int r; int err; do { char buf[READ_CHUNK]; r = safe_read(fd, buf, READ_CHUNK); if (r < 0) { err = -errno; cerr << "error while reading input" << std::endl; goto out; } bl.append(buf, r); } while (r > 0); err = 0; out: if (infile.size()) { close(fd); } return err; } template static int read_decode_json(const string& infile, T& t) { bufferlist bl; int ret = read_input(infile, bl); if (ret < 0) { cerr << "ERROR: failed to read input: " << cpp_strerror(-ret) << std::endl; return ret; } JSONParser p; if (!p.parse(bl.c_str(), bl.length())) { cout << "failed to parse JSON" << std::endl; return -EINVAL; } try { decode_json_obj(t, &p); } catch (JSONDecoder::err& e) { cout << "failed to decode JSON input: " << e.message << std::endl; return -EINVAL; } return 0; } template static int read_decode_json(const string& infile, T& t, K *k) { bufferlist bl; int ret = read_input(infile, bl); if (ret < 0) { cerr << "ERROR: failed to read input: " << cpp_strerror(-ret) << std::endl; return ret; } JSONParser p; if (!p.parse(bl.c_str(), bl.length())) { cout << "failed to parse JSON" << std::endl; return -EINVAL; } try { t.decode_json(&p, k); } catch (JSONDecoder::err& e) { cout << "failed to decode JSON input: " << e.message << std::endl; return -EINVAL; } return 0; } static int parse_date_str(const string& date_str, utime_t& ut) { uint64_t epoch = 0; uint64_t nsec = 0; if (!date_str.empty()) { int ret = utime_t::parse_date(date_str, &epoch, &nsec); if (ret < 0) { cerr << "ERROR: failed to parse date: " << date_str << std::endl; return -EINVAL; } } ut = utime_t(epoch, nsec); return 0; } template static bool decode_dump(const char *field_name, bufferlist& bl, Formatter *f) { T t; bufferlist::iterator iter = bl.begin(); try { ::decode(t, iter); } catch (buffer::error& err) { return false; } encode_json(field_name, t, f); return true; } static bool dump_string(const char *field_name, bufferlist& bl, Formatter *f) { string val; if (bl.length() > 0) { val.assign(bl.c_str()); } f->dump_string(field_name, val); return true; } void set_quota_info(RGWQuotaInfo& quota, int opt_cmd, int64_t max_size, int64_t max_objects, bool have_max_size, bool have_max_objects) { switch (opt_cmd) { case OPT_QUOTA_ENABLE: case OPT_GLOBAL_QUOTA_ENABLE: quota.enabled = true; // falling through on purpose case OPT_QUOTA_SET: case OPT_GLOBAL_QUOTA_SET: if (have_max_objects) { if (max_objects < 0) { quota.max_objects = -1; } else { quota.max_objects = max_objects; } } if (have_max_size) { if (max_size < 0) { quota.max_size = -1; } else { quota.max_size = rgw_rounded_kb(max_size) * 1024; } } break; case OPT_QUOTA_DISABLE: case OPT_GLOBAL_QUOTA_DISABLE: quota.enabled = false; break; } } int set_bucket_quota(RGWRados *store, int opt_cmd, const string& tenant_name, const string& bucket_name, int64_t max_size, int64_t max_objects, bool have_max_size, bool have_max_objects) { RGWBucketInfo bucket_info; map attrs; RGWObjectCtx obj_ctx(store); int r = store->get_bucket_info(obj_ctx, tenant_name, bucket_name, bucket_info, NULL, &attrs); if (r < 0) { cerr << "could not get bucket info for bucket=" << bucket_name << ": " << cpp_strerror(-r) << std::endl; return -r; } set_quota_info(bucket_info.quota, opt_cmd, max_size, max_objects, have_max_size, have_max_objects); r = store->put_bucket_instance_info(bucket_info, false, real_time(), &attrs); if (r < 0) { cerr << "ERROR: failed writing bucket instance info: " << cpp_strerror(-r) << std::endl; return -r; } return 0; } int set_user_bucket_quota(int opt_cmd, RGWUser& user, RGWUserAdminOpState& op_state, int64_t max_size, int64_t max_objects, bool have_max_size, bool have_max_objects) { RGWUserInfo& user_info = op_state.get_user_info(); set_quota_info(user_info.bucket_quota, opt_cmd, max_size, max_objects, have_max_size, have_max_objects); op_state.set_bucket_quota(user_info.bucket_quota); string err; int r = user.modify(op_state, &err); if (r < 0) { cerr << "ERROR: failed updating user info: " << cpp_strerror(-r) << ": " << err << std::endl; return -r; } return 0; } int set_user_quota(int opt_cmd, RGWUser& user, RGWUserAdminOpState& op_state, int64_t max_size, int64_t max_objects, bool have_max_size, bool have_max_objects) { RGWUserInfo& user_info = op_state.get_user_info(); set_quota_info(user_info.user_quota, opt_cmd, max_size, max_objects, have_max_size, have_max_objects); op_state.set_user_quota(user_info.user_quota); string err; int r = user.modify(op_state, &err); if (r < 0) { cerr << "ERROR: failed updating user info: " << cpp_strerror(-r) << ": " << err << std::endl; return -r; } return 0; } static bool bucket_object_check_filter(const string& name) { rgw_obj_key k; string ns; /* empty namespace */ return rgw_obj_key::oid_to_key_in_ns(name, &k, ns); } int check_min_obj_stripe_size(RGWRados *store, RGWBucketInfo& bucket_info, rgw_obj& obj, uint64_t min_stripe_size, bool *need_rewrite) { map attrs; uint64_t obj_size; RGWObjectCtx obj_ctx(store); RGWRados::Object op_target(store, bucket_info, obj_ctx, obj); RGWRados::Object::Read read_op(&op_target); read_op.params.attrs = &attrs; read_op.params.obj_size = &obj_size; int ret = read_op.prepare(); if (ret < 0) { lderr(store->ctx()) << "ERROR: failed to stat object, returned error: " << cpp_strerror(-ret) << dendl; return ret; } map::iterator iter; iter = attrs.find(RGW_ATTR_MANIFEST); if (iter == attrs.end()) { *need_rewrite = (obj_size >= min_stripe_size); return 0; } RGWObjManifest manifest; try { bufferlist& bl = iter->second; bufferlist::iterator biter = bl.begin(); ::decode(manifest, biter); } catch (buffer::error& err) { ldout(store->ctx(), 0) << "ERROR: failed to decode manifest" << dendl; return -EIO; } map& objs = manifest.get_explicit_objs(); map::iterator oiter; for (oiter = objs.begin(); oiter != objs.end(); ++oiter) { RGWObjManifestPart& part = oiter->second; if (part.size >= min_stripe_size) { *need_rewrite = true; return 0; } } *need_rewrite = false; return 0; } int check_obj_locator_underscore(RGWBucketInfo& bucket_info, rgw_obj& obj, rgw_obj_key& key, bool fix, bool remove_bad, Formatter *f) { f->open_object_section("object"); f->open_object_section("key"); f->dump_string("type", "head"); f->dump_string("name", key.name); f->dump_string("instance", key.instance); f->close_section(); string oid; string locator; get_obj_bucket_and_oid_loc(obj, oid, locator); f->dump_string("oid", oid); f->dump_string("locator", locator); RGWObjectCtx obj_ctx(store); RGWRados::Object op_target(store, bucket_info, obj_ctx, obj); RGWRados::Object::Read read_op(&op_target); int ret = read_op.prepare(); bool needs_fixing = (ret == -ENOENT); f->dump_bool("needs_fixing", needs_fixing); string status = (needs_fixing ? "needs_fixing" : "ok"); if ((needs_fixing || remove_bad) && fix) { ret = store->fix_head_obj_locator(bucket_info, needs_fixing, remove_bad, key); if (ret < 0) { cerr << "ERROR: fix_head_object_locator() returned ret=" << ret << std::endl; goto done; } status = "fixed"; } done: f->dump_string("status", status); f->close_section(); return 0; } int check_obj_tail_locator_underscore(RGWBucketInfo& bucket_info, rgw_obj& obj, rgw_obj_key& key, bool fix, Formatter *f) { f->open_object_section("object"); f->open_object_section("key"); f->dump_string("type", "tail"); f->dump_string("name", key.name); f->dump_string("instance", key.instance); f->close_section(); bool needs_fixing; string status; int ret = store->fix_tail_obj_locator(bucket_info, key, fix, &needs_fixing); if (ret < 0) { cerr << "ERROR: fix_tail_object_locator_underscore() returned ret=" << ret << std::endl; status = "failed"; } else { status = (needs_fixing && !fix ? "needs_fixing" : "ok"); } f->dump_bool("needs_fixing", needs_fixing); f->dump_string("status", status); f->close_section(); return 0; } int do_check_object_locator(const string& tenant_name, const string& bucket_name, bool fix, bool remove_bad, Formatter *f) { if (remove_bad && !fix) { cerr << "ERROR: can't have remove_bad specified without fix" << std::endl; return -EINVAL; } RGWBucketInfo bucket_info; rgw_bucket bucket; string bucket_id; f->open_object_section("bucket"); f->dump_string("bucket", bucket_name); int ret = init_bucket(tenant_name, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return ret; } bool truncated; int count = 0; int max_entries = 1000; string prefix; string delim; vector result; map common_prefixes; string ns; RGWRados::Bucket target(store, bucket_info); RGWRados::Bucket::List list_op(&target); string marker; list_op.params.prefix = prefix; list_op.params.delim = delim; list_op.params.marker = rgw_obj_key(marker); list_op.params.ns = ns; list_op.params.enforce_ns = true; list_op.params.list_versions = true; f->open_array_section("check_objects"); do { ret = list_op.list_objects(max_entries - count, &result, &common_prefixes, &truncated); if (ret < 0) { cerr << "ERROR: store->list_objects(): " << cpp_strerror(-ret) << std::endl; return -ret; } count += result.size(); for (vector::iterator iter = result.begin(); iter != result.end(); ++iter) { rgw_obj_key key = iter->key; rgw_obj obj(bucket, key); if (key.name[0] == '_') { ret = check_obj_locator_underscore(bucket_info, obj, key, fix, remove_bad, f); if (ret >= 0) { ret = check_obj_tail_locator_underscore(bucket_info, obj, key, fix, f); } } } f->flush(cout); } while (truncated && count < max_entries); f->close_section(); f->close_section(); f->flush(cout); return 0; } int set_bucket_sync_enabled(RGWRados *store, int opt_cmd, const string& tenant_name, const string& bucket_name) { RGWBucketInfo bucket_info; map attrs; RGWObjectCtx obj_ctx(store); int r = store->get_bucket_info(obj_ctx, tenant_name, bucket_name, bucket_info, NULL, &attrs); if (r < 0) { cerr << "could not get bucket info for bucket=" << bucket_name << ": " << cpp_strerror(-r) << std::endl; return -r; } if (opt_cmd == OPT_BUCKET_SYNC_ENABLE) { bucket_info.flags &= ~BUCKET_DATASYNC_DISABLED; } else if (opt_cmd == OPT_BUCKET_SYNC_DISABLE) { bucket_info.flags |= BUCKET_DATASYNC_DISABLED; } r = store->put_bucket_instance_info(bucket_info, false, real_time(), &attrs); if (r < 0) { cerr << "ERROR: failed writing bucket instance info: " << cpp_strerror(-r) << std::endl; return -r; } int shards_num = bucket_info.num_shards? bucket_info.num_shards : 1; int shard_id = bucket_info.num_shards? 0 : -1; if (opt_cmd == OPT_BUCKET_SYNC_DISABLE) { r = store->stop_bi_log_entries(bucket_info, -1); if (r < 0) { lderr(store->ctx()) << "ERROR: failed writing stop bilog" << dendl; return r; } } else { r = store->resync_bi_log_entries(bucket_info, -1); if (r < 0) { lderr(store->ctx()) << "ERROR: failed writing resync bilog" << dendl; return r; } } for (int i = 0; i < shards_num; ++i, ++shard_id) { r = store->data_log->add_entry(bucket_info.bucket, shard_id); if (r < 0) { lderr(store->ctx()) << "ERROR: failed writing data log" << dendl; return r; } } return 0; } /// search for a matching zone/zonegroup id and return a connection if found static boost::optional get_remote_conn(RGWRados *store, const RGWZoneGroup& zonegroup, const std::string& remote) { boost::optional conn; if (remote == zonegroup.get_id()) { conn.emplace(store->ctx(), store, remote, zonegroup.endpoints); } else { for (const auto& z : zonegroup.zones) { const auto& zone = z.second; if (remote == zone.id) { conn.emplace(store->ctx(), store, remote, zone.endpoints); break; } } } return conn; } /// search each zonegroup for a connection static boost::optional get_remote_conn(RGWRados *store, const RGWPeriodMap& period_map, const std::string& remote) { boost::optional conn; for (const auto& zg : period_map.zonegroups) { conn = get_remote_conn(store, zg.second, remote); if (conn) { break; } } return conn; } // we expect a very small response static constexpr size_t MAX_REST_RESPONSE = 128 * 1024; static int send_to_remote_gateway(RGWRESTConn* conn, req_info& info, bufferlist& in_data, JSONParser& parser) { if (!conn) { return -EINVAL; } ceph::bufferlist response; rgw_user user; int ret = conn->forward(user, info, nullptr, MAX_REST_RESPONSE, &in_data, &response); int parse_ret = parser.parse(response.c_str(), response.length()); if (parse_ret < 0) { cerr << "failed to parse response" << std::endl; return parse_ret; } return ret; } static int send_to_url(const string& url, const string& access, const string& secret, req_info& info, bufferlist& in_data, JSONParser& parser) { if (access.empty() || secret.empty()) { cerr << "An --access-key and --secret must be provided with --url." << std::endl; return -EINVAL; } RGWAccessKey key; key.id = access; key.key = secret; param_vec_t params; RGWRESTSimpleRequest req(g_ceph_context, url, NULL, ¶ms); bufferlist response; int ret = req.forward_request(key, info, MAX_REST_RESPONSE, &in_data, &response); int parse_ret = parser.parse(response.c_str(), response.length()); if (parse_ret < 0) { cout << "failed to parse response" << std::endl; return parse_ret; } return ret; } static int send_to_remote_or_url(RGWRESTConn *conn, const string& url, const string& access, const string& secret, req_info& info, bufferlist& in_data, JSONParser& parser) { if (url.empty()) { return send_to_remote_gateway(conn, info, in_data, parser); } return send_to_url(url, access, secret, info, in_data, parser); } static int commit_period(RGWRealm& realm, RGWPeriod& period, string remote, const string& url, const string& access, const string& secret, bool force) { const string& master_zone = period.get_master_zone(); if (master_zone.empty()) { cerr << "cannot commit period: period does not have a master zone of a master zonegroup" << std::endl; return -EINVAL; } // are we the period's master zone? if (store->get_zone_params().get_id() == master_zone) { // read the current period RGWPeriod current_period; int ret = current_period.init(g_ceph_context, store, realm.get_id()); if (ret < 0) { cerr << "Error initializing current period: " << cpp_strerror(-ret) << std::endl; return ret; } // the master zone can commit locally ret = period.commit(realm, current_period, cerr, force); if (ret < 0) { cerr << "failed to commit period: " << cpp_strerror(-ret) << std::endl; } return ret; } if (remote.empty() && url.empty()) { // use the new master zone's connection remote = master_zone; cout << "Sending period to new master zone " << remote << std::endl; } boost::optional conn; RGWRESTConn *remote_conn = nullptr; if (!remote.empty()) { conn = get_remote_conn(store, period.get_map(), remote); if (!conn) { cerr << "failed to find a zone or zonegroup for remote " << remote << std::endl; return -ENOENT; } remote_conn = &*conn; } // push period to the master with an empty period id period.set_id(""); RGWEnv env; req_info info(g_ceph_context, &env); info.method = "POST"; info.request_uri = "/admin/realm/period"; // json format into a bufferlist JSONFormatter jf(false); encode_json("period", period, &jf); bufferlist bl; jf.flush(bl); JSONParser p; int ret = send_to_remote_or_url(remote_conn, url, access, secret, info, bl, p); if (ret < 0) { cerr << "request failed: " << cpp_strerror(-ret) << std::endl; // did we parse an error message? auto message = p.find_obj("Message"); if (message) { cerr << "Reason: " << message->get_data() << std::endl; } return ret; } // decode the response and store it back try { decode_json_obj(period, &p); } catch (JSONDecoder::err& e) { cout << "failed to decode JSON input: " << e.message << std::endl; return -EINVAL; } if (period.get_id().empty()) { cerr << "Period commit got back an empty period id" << std::endl; return -EINVAL; } // the master zone gave us back the period that it committed, so it's // safe to save it as our latest epoch ret = period.store_info(false); if (ret < 0) { cerr << "Error storing committed period " << period.get_id() << ": " << cpp_strerror(ret) << std::endl; return ret; } ret = period.set_latest_epoch(period.get_epoch()); if (ret < 0) { cerr << "Error updating period epoch: " << cpp_strerror(ret) << std::endl; return ret; } ret = period.reflect(); if (ret < 0) { cerr << "Error updating local objects: " << cpp_strerror(ret) << std::endl; return ret; } realm.notify_new_period(period); return ret; } static int update_period(const string& realm_id, const string& realm_name, const string& period_id, const string& period_epoch, bool commit, const string& remote, const string& url, const string& access, const string& secret, Formatter *formatter, bool force) { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0 ) { cerr << "Error initializing realm " << cpp_strerror(-ret) << std::endl; return ret; } epoch_t epoch = 0; if (!period_epoch.empty()) { epoch = atoi(period_epoch.c_str()); } RGWPeriod period(period_id, epoch); ret = period.init(g_ceph_context, store, realm.get_id()); if (ret < 0) { cerr << "period init failed: " << cpp_strerror(-ret) << std::endl; return ret; } period.fork(); ret = period.update(); if(ret < 0) { // Dropping the error message here, as both the ret codes were handled in // period.update() return ret; } ret = period.store_info(false); if (ret < 0) { cerr << "failed to store period: " << cpp_strerror(-ret) << std::endl; return ret; } if (commit) { ret = commit_period(realm, period, remote, url, access, secret, force); if (ret < 0) { cerr << "failed to commit period: " << cpp_strerror(-ret) << std::endl; return ret; } } encode_json("period", period, formatter); formatter->flush(cout); cout << std::endl; return 0; } static int init_bucket_for_sync(const string& tenant, const string& bucket_name, const string& bucket_id, rgw_bucket& bucket) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return ret; } return 0; } static int do_period_pull(RGWRESTConn *remote_conn, const string& url, const string& access_key, const string& secret_key, const string& realm_id, const string& realm_name, const string& period_id, const string& period_epoch, RGWPeriod *period) { RGWEnv env; req_info info(g_ceph_context, &env); info.method = "GET"; info.request_uri = "/admin/realm/period"; map ¶ms = info.args.get_params(); if (!realm_id.empty()) params["realm_id"] = realm_id; if (!realm_name.empty()) params["realm_name"] = realm_name; if (!period_id.empty()) params["period_id"] = period_id; if (!period_epoch.empty()) params["epoch"] = period_epoch; bufferlist bl; JSONParser p; int ret = send_to_remote_or_url(remote_conn, url, access_key, secret_key, info, bl, p); if (ret < 0) { cerr << "request failed: " << cpp_strerror(-ret) << std::endl; return ret; } ret = period->init(g_ceph_context, store, false); if (ret < 0) { cerr << "faile to init period " << cpp_strerror(-ret) << std::endl; return ret; } try { decode_json_obj(*period, &p); } catch (JSONDecoder::err& e) { cout << "failed to decode JSON input: " << e.message << std::endl; return -EINVAL; } ret = period->store_info(false); if (ret < 0) { cerr << "Error storing period " << period->get_id() << ": " << cpp_strerror(ret) << std::endl; } // store latest epoch (ignore errors) period->update_latest_epoch(period->get_epoch()); return 0; } static int read_current_period_id(RGWRados* store, const std::string& realm_id, const std::string& realm_name, std::string* period_id) { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0) { std::cerr << "failed to read realm: " << cpp_strerror(-ret) << std::endl; return ret; } *period_id = realm.get_current_period(); return 0; } void flush_ss(stringstream& ss, list& l) { if (!ss.str().empty()) { l.push_back(ss.str()); } ss.str(""); } stringstream& push_ss(stringstream& ss, list& l, int tab = 0) { flush_ss(ss, l); if (tab > 0) { ss << setw(tab) << "" << setw(1); } return ss; } static void get_md_sync_status(list& status) { RGWMetaSyncStatusManager sync(store, store->get_async_rados()); int ret = sync.init(); if (ret < 0) { status.push_back(string("failed to retrieve sync info: sync.init() failed: ") + cpp_strerror(-ret)); return; } rgw_meta_sync_status sync_status; ret = sync.read_sync_status(&sync_status); if (ret < 0) { status.push_back(string("failed to read sync status: ") + cpp_strerror(-ret)); return; } string status_str; switch (sync_status.sync_info.state) { case rgw_meta_sync_info::StateInit: status_str = "init"; break; case rgw_meta_sync_info::StateBuildingFullSyncMaps: status_str = "preparing for full sync"; break; case rgw_meta_sync_info::StateSync: status_str = "syncing"; break; default: status_str = "unknown"; } status.push_back(status_str); uint64_t full_total = 0; uint64_t full_complete = 0; int num_full = 0; int num_inc = 0; int total_shards = 0; set shards_behind_set; for (auto marker_iter : sync_status.sync_markers) { full_total += marker_iter.second.total_entries; total_shards++; int shard_id = marker_iter.first; if (marker_iter.second.state == rgw_meta_sync_marker::SyncState::FullSync) { num_full++; full_complete += marker_iter.second.pos; shards_behind_set.insert(shard_id); } else { full_complete += marker_iter.second.total_entries; } if (marker_iter.second.state == rgw_meta_sync_marker::SyncState::IncrementalSync) { num_inc++; } } stringstream ss; push_ss(ss, status) << "full sync: " << num_full << "/" << total_shards << " shards"; if (num_full > 0) { push_ss(ss, status) << "full sync: " << full_total - full_complete << " entries to sync"; } push_ss(ss, status) << "incremental sync: " << num_inc << "/" << total_shards << " shards"; rgw_mdlog_info log_info; ret = sync.read_log_info(&log_info); if (ret < 0) { status.push_back(string("failed to fetch local sync status: ") + cpp_strerror(-ret)); return; } map master_shards_info; string master_period = store->get_current_period_id(); ret = sync.read_master_log_shards_info(master_period, &master_shards_info); if (ret < 0) { status.push_back(string("failed to fetch master sync status: ") + cpp_strerror(-ret)); return; } map shards_behind; if (sync_status.sync_info.period != master_period) { status.push_back(string("master is on a different period: master_period=" + master_period + " local_period=" + sync_status.sync_info.period)); } else { for (auto local_iter : sync_status.sync_markers) { int shard_id = local_iter.first; auto iter = master_shards_info.find(shard_id); if (iter == master_shards_info.end()) { /* huh? */ derr << "ERROR: could not find remote sync shard status for shard_id=" << shard_id << dendl; continue; } auto master_marker = iter->second.marker; if (local_iter.second.state == rgw_meta_sync_marker::SyncState::IncrementalSync && master_marker > local_iter.second.marker) { shards_behind[shard_id] = local_iter.second.marker; shards_behind_set.insert(shard_id); } } } int total_behind = shards_behind.size() + (sync_status.sync_info.num_shards - num_inc); if (total_behind == 0) { push_ss(ss, status) << "metadata is caught up with master"; } else { push_ss(ss, status) << "metadata is behind on " << total_behind << " shards"; push_ss(ss, status) << "behind shards: " << "[" << shards_behind_set << "]"; map master_pos; ret = sync.read_master_log_shards_next(sync_status.sync_info.period, shards_behind, &master_pos); if (ret < 0) { derr << "ERROR: failed to fetch master next positions (" << cpp_strerror(-ret) << ")" << dendl; } else { ceph::real_time oldest; for (auto iter : master_pos) { rgw_mdlog_shard_data& shard_data = iter.second; if (!shard_data.entries.empty()) { rgw_mdlog_entry& entry = shard_data.entries.front(); if (ceph::real_clock::is_zero(oldest)) { oldest = entry.timestamp; } else if (!ceph::real_clock::is_zero(entry.timestamp) && entry.timestamp < oldest) { oldest = entry.timestamp; } } } if (!ceph::real_clock::is_zero(oldest)) { push_ss(ss, status) << "oldest incremental change not applied: " << oldest; } } } flush_ss(ss, status); } static void get_data_sync_status(const string& source_zone, list& status, int tab) { stringstream ss; auto ziter = store->zone_by_id.find(source_zone); if (ziter == store->zone_by_id.end()) { push_ss(ss, status, tab) << string("zone not found"); flush_ss(ss, status); return; } RGWZone& sz = ziter->second; if (!store->zone_syncs_from(store->get_zone(), sz)) { push_ss(ss, status, tab) << string("not syncing from zone"); flush_ss(ss, status); return; } RGWDataSyncStatusManager sync(store, store->get_async_rados(), source_zone); int ret = sync.init(); if (ret < 0) { push_ss(ss, status, tab) << string("failed to retrieve sync info: ") + cpp_strerror(-ret); flush_ss(ss, status); return; } rgw_data_sync_status sync_status; ret = sync.read_sync_status(&sync_status); if (ret < 0 && ret != -ENOENT) { push_ss(ss, status, tab) << string("failed read sync status: ") + cpp_strerror(-ret); return; } set recovering_shards; ret = sync.read_recovering_shards(sync_status.sync_info.num_shards, recovering_shards); if (ret < 0 && ret != ENOENT) { push_ss(ss, status, tab) << string("failed read recovering shards: ") + cpp_strerror(-ret); return; } string status_str; switch (sync_status.sync_info.state) { case rgw_data_sync_info::StateInit: status_str = "init"; break; case rgw_data_sync_info::StateBuildingFullSyncMaps: status_str = "preparing for full sync"; break; case rgw_data_sync_info::StateSync: status_str = "syncing"; break; default: status_str = "unknown"; } push_ss(ss, status, tab) << status_str; uint64_t full_total = 0; uint64_t full_complete = 0; int num_full = 0; int num_inc = 0; int total_shards = 0; set shards_behind_set; for (auto marker_iter : sync_status.sync_markers) { int shard_id = marker_iter.first; full_total += marker_iter.second.total_entries; total_shards++; if (marker_iter.second.state == rgw_data_sync_marker::SyncState::FullSync) { num_full++; full_complete += marker_iter.second.pos; shards_behind_set.insert(shard_id); } else { full_complete += marker_iter.second.total_entries; } if (marker_iter.second.state == rgw_data_sync_marker::SyncState::IncrementalSync) { num_inc++; } } push_ss(ss, status, tab) << "full sync: " << num_full << "/" << total_shards << " shards"; if (num_full > 0) { push_ss(ss, status, tab) << "full sync: " << full_total - full_complete << " buckets to sync"; } push_ss(ss, status, tab) << "incremental sync: " << num_inc << "/" << total_shards << " shards"; rgw_datalog_info log_info; ret = sync.read_log_info(&log_info); if (ret < 0) { push_ss(ss, status, tab) << string("failed to fetch local sync status: ") + cpp_strerror(-ret); return; } map source_shards_info; ret = sync.read_source_log_shards_info(&source_shards_info); if (ret < 0) { push_ss(ss, status, tab) << string("failed to fetch source sync status: ") + cpp_strerror(-ret); return; } map shards_behind; for (auto local_iter : sync_status.sync_markers) { int shard_id = local_iter.first; auto iter = source_shards_info.find(shard_id); if (iter == source_shards_info.end()) { /* huh? */ derr << "ERROR: could not find remote sync shard status for shard_id=" << shard_id << dendl; continue; } auto master_marker = iter->second.marker; if (local_iter.second.state == rgw_data_sync_marker::SyncState::IncrementalSync && master_marker > local_iter.second.marker) { shards_behind[shard_id] = local_iter.second.marker; shards_behind_set.insert(shard_id); } } int total_behind = shards_behind.size() + (sync_status.sync_info.num_shards - num_inc); int total_recovering = recovering_shards.size(); if (total_behind == 0 && total_recovering == 0) { push_ss(ss, status, tab) << "data is caught up with source"; } else if (total_behind > 0) { push_ss(ss, status, tab) << "data is behind on " << total_behind << " shards"; push_ss(ss, status, tab) << "behind shards: " << "[" << shards_behind_set << "]" ; map master_pos; ret = sync.read_source_log_shards_next(shards_behind, &master_pos); if (ret < 0) { derr << "ERROR: failed to fetch next positions (" << cpp_strerror(-ret) << ")" << dendl; } else { ceph::real_time oldest; for (auto iter : master_pos) { rgw_datalog_shard_data& shard_data = iter.second; if (!shard_data.entries.empty()) { rgw_datalog_entry& entry = shard_data.entries.front(); if (ceph::real_clock::is_zero(oldest)) { oldest = entry.timestamp; } else if (!ceph::real_clock::is_zero(entry.timestamp) && entry.timestamp < oldest) { oldest = entry.timestamp; } } } if (!ceph::real_clock::is_zero(oldest)) { push_ss(ss, status, tab) << "oldest incremental change not applied: " << oldest; } } } if (total_recovering > 0) { push_ss(ss, status, tab) << total_recovering << " shards are recovering"; push_ss(ss, status, tab) << "recovering shards: " << "[" << recovering_shards << "]"; } flush_ss(ss, status); } static void tab_dump(const string& header, int width, const list& entries) { string s = header; for (auto e : entries) { cout << std::setw(width) << s << std::setw(1) << " " << e << std::endl; s.clear(); } } static void sync_status(Formatter *formatter) { RGWRealm& realm = store->realm; RGWZoneGroup& zonegroup = store->get_zonegroup(); RGWZone& zone = store->get_zone(); int width = 15; cout << std::setw(width) << "realm" << std::setw(1) << " " << realm.get_id() << " (" << realm.get_name() << ")" << std::endl; cout << std::setw(width) << "zonegroup" << std::setw(1) << " " << zonegroup.get_id() << " (" << zonegroup.get_name() << ")" << std::endl; cout << std::setw(width) << "zone" << std::setw(1) << " " << zone.id << " (" << zone.name << ")" << std::endl; list md_status; if (store->is_meta_master()) { md_status.push_back("no sync (zone is master)"); } else { get_md_sync_status(md_status); } tab_dump("metadata sync", width, md_status); list data_status; for (auto iter : store->zone_conn_map) { const string& source_id = iter.first; string source_str = "source: "; string s = source_str + source_id; auto siter = store->zone_by_id.find(source_id); if (siter != store->zone_by_id.end()) { s += string(" (") + siter->second.name + ")"; } data_status.push_back(s); get_data_sync_status(source_id, data_status, source_str.size()); } tab_dump("data sync", width, data_status); } struct indented { int w; // indent width boost::string_view header; indented(int w, boost::string_view header = "") : w(w), header(header) {} }; std::ostream& operator<<(std::ostream& out, const indented& h) { return out << std::setw(h.w) << h.header << std::setw(1) << ' '; } static int remote_bilog_markers(RGWRados *store, const RGWZone& source, RGWRESTConn *conn, const RGWBucketInfo& info, BucketIndexShardsManager *markers) { const auto instance_key = info.bucket.get_key(); const rgw_http_param_pair params[] = { { "type" , "bucket-index" }, { "bucket-instance", instance_key.c_str() }, { "info" , nullptr }, { nullptr, nullptr } }; rgw_bucket_index_marker_info result; int r = conn->get_json_resource("/admin/log/", params, result); if (r < 0) { lderr(store->ctx()) << "failed to fetch remote log markers: " << cpp_strerror(r) << dendl; return r; } r = markers->from_string(result.max_marker, -1); if (r < 0) { lderr(store->ctx()) << "failed to decode remote log markers" << dendl; return r; } return 0; } static int bucket_source_sync_status(RGWRados *store, const RGWZone& zone, const RGWZone& source, RGWRESTConn *conn, const RGWBucketInfo& bucket_info, int width, std::ostream& out) { out << indented{width, "source zone"} << source.id << " (" << source.name << ")\n"; // syncing from this zone? if (!zone.syncs_from(source.id)) { out << indented{width} << "not in sync_from\n"; return 0; } std::vector status; int r = rgw_bucket_sync_status(store, source.id, bucket_info, &status); if (r < 0) { lderr(store->ctx()) << "failed to read bucket sync status: " << cpp_strerror(r) << dendl; return r; } int num_full = 0; int num_inc = 0; uint64_t full_complete = 0; const int total_shards = status.size(); using BucketSyncState = rgw_bucket_shard_sync_info::SyncState; for (size_t shard_id = 0; shard_id < total_shards; shard_id++) { auto& m = status[shard_id]; if (m.state == BucketSyncState::StateFullSync) { num_full++; full_complete += m.full_marker.count; } else if (m.state == BucketSyncState::StateIncrementalSync) { num_inc++; } } out << indented{width} << "full sync: " << num_full << "/" << total_shards << " shards\n"; if (num_full > 0) { out << indented{width} << "full sync: " << full_complete << " objects completed\n"; } out << indented{width} << "incremental sync: " << num_inc << "/" << total_shards << " shards\n"; BucketIndexShardsManager remote_markers; r = remote_bilog_markers(store, source, conn, bucket_info, &remote_markers); if (r < 0) { lderr(store->ctx()) << "failed to read remote log: " << cpp_strerror(r) << dendl; return r; } std::set shards_behind; for (auto& r : remote_markers.get()) { auto shard_id = r.first; auto& m = status[shard_id]; if (r.second.empty()) { continue; // empty bucket index shard } auto pos = BucketIndexShardsManager::get_shard_marker(m.inc_marker.position); if (m.state != BucketSyncState::StateIncrementalSync || pos != r.second) { shards_behind.insert(shard_id); } } if (shards_behind.empty()) { out << indented{width} << "bucket is caught up with source\n"; } else { out << indented{width} << "bucket is behind on " << shards_behind.size() << " shards\n"; out << indented{width} << "behind shards: [" << shards_behind << "]\n" ; } return 0; } static int bucket_sync_status(RGWRados *store, const RGWBucketInfo& info, const std::string& source_zone_id, std::ostream& out) { RGWRealm& realm = store->realm; RGWZoneGroup& zonegroup = store->get_zonegroup(); RGWZone& zone = store->get_zone(); constexpr int width = 15; out << indented{width, "realm"} << realm.get_id() << " (" << realm.get_name() << ")\n"; out << indented{width, "zonegroup"} << zonegroup.get_id() << " (" << zonegroup.get_name() << ")\n"; out << indented{width, "zone"} << zone.id << " (" << zone.name << ")\n"; out << indented{width, "bucket"} << info.bucket << "\n\n"; if (!info.datasync_flag_enabled()) { out << "Sync is disabled for bucket " << info.bucket.name << '\n'; return 0; } if (!source_zone_id.empty()) { auto z = zonegroup.zones.find(source_zone_id); if (z == zonegroup.zones.end()) { lderr(store->ctx()) << "Source zone not found in zonegroup " << zonegroup.get_name() << dendl; return -EINVAL; } auto c = store->zone_conn_map.find(source_zone_id); if (c == store->zone_conn_map.end()) { lderr(store->ctx()) << "No connection to zone " << z->second.name << dendl; return -EINVAL; } return bucket_source_sync_status(store, zone, z->second, c->second, info, width, out); } for (const auto& z : zonegroup.zones) { auto c = store->zone_conn_map.find(z.second.id); if (c != store->zone_conn_map.end()) { bucket_source_sync_status(store, zone, z.second, c->second, info, width, out); } } return 0; } static void parse_tier_config_param(const string& s, map& out) { list confs; get_str_list(s, ",", confs); for (auto c : confs) { ssize_t pos = c.find("="); if (pos < 0) { out[c] = ""; } else { out[c.substr(0, pos)] = c.substr(pos + 1); } } } int check_reshard_bucket_params(RGWRados *store, const string& bucket_name, const string& tenant, const string& bucket_id, bool num_shards_specified, int num_shards, int yes_i_really_mean_it, rgw_bucket& bucket, RGWBucketInfo& bucket_info, map& attrs) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return -EINVAL; } if (!num_shards_specified) { cerr << "ERROR: --num-shards not specified" << std::endl; return -EINVAL; } if (num_shards > (int)store->get_max_bucket_shards()) { cerr << "ERROR: num_shards too high, max value: " << store->get_max_bucket_shards() << std::endl; return -EINVAL; } int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket, &attrs); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } int num_source_shards = (bucket_info.num_shards > 0 ? bucket_info.num_shards : 1); if (num_shards <= num_source_shards && !yes_i_really_mean_it) { cerr << "num shards is less or equal to current shards count" << std::endl << "do you really mean it? (requires --yes-i-really-mean-it)" << std::endl; return -EINVAL; } return 0; } int create_new_bucket_instance(RGWRados *store, int new_num_shards, const RGWBucketInfo& bucket_info, map& attrs, RGWBucketInfo& new_bucket_info) { store->create_bucket_id(&new_bucket_info.bucket.bucket_id); new_bucket_info.bucket.oid.clear(); new_bucket_info.num_shards = new_num_shards; new_bucket_info.objv_tracker.clear(); int ret = store->init_bucket_index(new_bucket_info, new_bucket_info.num_shards); if (ret < 0) { cerr << "ERROR: failed to init new bucket indexes: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = store->put_bucket_instance_info(new_bucket_info, true, real_time(), &attrs); if (ret < 0) { cerr << "ERROR: failed to store new bucket instance info: " << cpp_strerror(-ret) << std::endl; return -ret; } return 0; } static int trim_sync_error_log(int shard_id, const ceph::real_time& start_time, const ceph::real_time& end_time, const string& start_marker, const string& end_marker, int delay_ms) { auto oid = RGWSyncErrorLogger::get_shard_oid(RGW_SYNC_ERROR_LOG_SHARD_PREFIX, shard_id); // call cls_log_trim() until it returns -ENODATA for (;;) { int ret = store->time_log_trim(oid, start_time, end_time, start_marker, end_marker); if (ret == -ENODATA) { return 0; } if (ret < 0) { return ret; } if (delay_ms) { std::this_thread::sleep_for(std::chrono::milliseconds(delay_ms)); } } // unreachable } int main(int argc, const char **argv) { vector args; argv_to_vec(argc, (const char **)argv, args); env_to_vec(args); auto cct = global_init(NULL, args, CEPH_ENTITY_TYPE_CLIENT, CODE_ENVIRONMENT_UTILITY, 0); // for region -> zonegroup conversion (must happen before common_init_finish()) if (!g_conf->rgw_region.empty() && g_conf->rgw_zonegroup.empty()) { g_conf->set_val_or_die("rgw_zonegroup", g_conf->rgw_region.c_str()); } common_init_finish(g_ceph_context); rgw_user user_id; string tenant; std::string access_key, secret_key, user_email, display_name; std::string bucket_name, pool_name, object; rgw_pool pool; std::string date, subuser, access, format; std::string start_date, end_date; std::string key_type_str; std::string period_id, period_epoch, remote, url; std::string master_zonegroup, master_zone; std::string realm_name, realm_id, realm_new_name; std::string zone_name, zone_id, zone_new_name; std::string zonegroup_name, zonegroup_id, zonegroup_new_name; std::string api_name; std::string role_name, path, assume_role_doc, policy_name, perm_policy_doc, path_prefix; list endpoints; int tmp_int; int sync_from_all_specified = false; bool sync_from_all = false; list sync_from; list sync_from_rm; std::string master_url; int is_master_int; int set_default = 0; bool is_master = false; bool is_master_set = false; int read_only_int; bool read_only = false; int is_read_only_set = false; int commit = false; int staging = false; int key_type = KEY_TYPE_UNDEFINED; rgw_bucket bucket; uint32_t perm_mask = 0; RGWUserInfo info; int opt_cmd = OPT_NO_CMD; bool need_more; int gen_access_key = 0; int gen_secret_key = 0; bool set_perm = false; bool set_temp_url_key = false; map temp_url_keys; string bucket_id; Formatter *formatter = NULL; int purge_data = false; int pretty_format = false; int show_log_entries = true; int show_log_sum = true; int skip_zero_entries = false; // log show int purge_keys = false; int yes_i_really_mean_it = false; int delete_child_objects = false; int fix = false; int remove_bad = false; int check_head_obj_locator = false; int max_buckets = -1; bool max_buckets_specified = false; map categories; string caps; int check_objects = false; RGWUserAdminOpState user_op; RGWBucketAdminOpState bucket_op; string infile; string metadata_key; RGWObjVersionTracker objv_tracker; string marker; string start_marker; string end_marker; int max_entries = -1; bool max_entries_specified = false; int admin = false; bool admin_specified = false; int system = false; bool system_specified = false; int shard_id = -1; bool specified_shard_id = false; string daemon_id; bool specified_daemon_id = false; string client_id; string op_id; string state_str; string replica_log_type_str; ReplicaLogType replica_log_type = ReplicaLog_Invalid; string op_mask_str; string quota_scope; string object_version; string placement_id; list tags; list tags_add; list tags_rm; int64_t max_objects = -1; int64_t max_size = -1; bool have_max_objects = false; bool have_max_size = false; int include_all = false; int sync_stats = false; int reset_stats = false; int bypass_gc = false; int warnings_only = false; int inconsistent_index = false; int verbose = false; int extra_info = false; uint64_t min_rewrite_size = 4 * 1024 * 1024; uint64_t max_rewrite_size = ULLONG_MAX; uint64_t min_rewrite_stripe_size = 0; BIIndexType bi_index_type = PlainIdx; string job_id; int num_shards = 0; bool num_shards_specified = false; int max_concurrent_ios = 32; uint64_t orphan_stale_secs = (24 * 3600); std::string val; std::ostringstream errs; string err; long long tmp = 0; string source_zone_name; string source_zone; /* zone id */ string tier_type; bool tier_type_specified = false; map tier_config_add; map tier_config_rm; boost::optional index_pool; boost::optional data_pool; boost::optional data_extra_pool; RGWBucketIndexType placement_index_type = RGWBIType_Normal; bool index_type_specified = false; boost::optional compression_type; int trim_delay_ms = 0; for (std::vector::iterator i = args.begin(); i != args.end(); ) { if (ceph_argparse_double_dash(args, i)) { break; } else if (ceph_argparse_flag(args, i, "-h", "--help", (char*)NULL)) { usage(); assert(false); } else if (ceph_argparse_witharg(args, i, &val, "-i", "--uid", (char*)NULL)) { user_id.from_str(val); } else if (ceph_argparse_witharg(args, i, &val, "--tenant", (char*)NULL)) { tenant = val; } else if (ceph_argparse_witharg(args, i, &val, "--access-key", (char*)NULL)) { access_key = val; } else if (ceph_argparse_witharg(args, i, &val, "--subuser", (char*)NULL)) { subuser = val; } else if (ceph_argparse_witharg(args, i, &val, "--secret", "--secret-key", (char*)NULL)) { secret_key = val; } else if (ceph_argparse_witharg(args, i, &val, "-e", "--email", (char*)NULL)) { user_email = val; user_op.user_email_specified=true; } else if (ceph_argparse_witharg(args, i, &val, "-n", "--display-name", (char*)NULL)) { display_name = val; } else if (ceph_argparse_witharg(args, i, &val, "-b", "--bucket", (char*)NULL)) { bucket_name = val; } else if (ceph_argparse_witharg(args, i, &val, "-p", "--pool", (char*)NULL)) { pool_name = val; pool = rgw_pool(pool_name); } else if (ceph_argparse_witharg(args, i, &val, "-o", "--object", (char*)NULL)) { object = val; } else if (ceph_argparse_witharg(args, i, &val, "--object-version", (char*)NULL)) { object_version = val; } else if (ceph_argparse_witharg(args, i, &val, "--client-id", (char*)NULL)) { client_id = val; } else if (ceph_argparse_witharg(args, i, &val, "--op-id", (char*)NULL)) { op_id = val; } else if (ceph_argparse_witharg(args, i, &val, "--state", (char*)NULL)) { state_str = val; } else if (ceph_argparse_witharg(args, i, &val, "--op-mask", (char*)NULL)) { op_mask_str = val; } else if (ceph_argparse_witharg(args, i, &val, "--key-type", (char*)NULL)) { key_type_str = val; if (key_type_str.compare("swift") == 0) { key_type = KEY_TYPE_SWIFT; } else if (key_type_str.compare("s3") == 0) { key_type = KEY_TYPE_S3; } else { cerr << "bad key type: " << key_type_str << std::endl; usage(); assert(false); } } else if (ceph_argparse_witharg(args, i, &val, "--job-id", (char*)NULL)) { job_id = val; } else if (ceph_argparse_binary_flag(args, i, &gen_access_key, NULL, "--gen-access-key", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &gen_secret_key, NULL, "--gen-secret", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &show_log_entries, NULL, "--show_log_entries", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &show_log_sum, NULL, "--show_log_sum", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &skip_zero_entries, NULL, "--skip_zero_entries", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &admin, NULL, "--admin", (char*)NULL)) { admin_specified = true; } else if (ceph_argparse_binary_flag(args, i, &system, NULL, "--system", (char*)NULL)) { system_specified = true; } else if (ceph_argparse_binary_flag(args, i, &verbose, NULL, "--verbose", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &staging, NULL, "--staging", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &commit, NULL, "--commit", (char*)NULL)) { // do nothing } else if (ceph_argparse_witharg(args, i, &tmp, errs, "-a", "--auth-uid", (char*)NULL)) { if (!errs.str().empty()) { cerr << errs.str() << std::endl; exit(EXIT_FAILURE); } } else if (ceph_argparse_witharg(args, i, &val, "--min-rewrite-size", (char*)NULL)) { min_rewrite_size = (uint64_t)atoll(val.c_str()); } else if (ceph_argparse_witharg(args, i, &val, "--max-rewrite-size", (char*)NULL)) { max_rewrite_size = (uint64_t)atoll(val.c_str()); } else if (ceph_argparse_witharg(args, i, &val, "--min-rewrite-stripe-size", (char*)NULL)) { min_rewrite_stripe_size = (uint64_t)atoll(val.c_str()); } else if (ceph_argparse_witharg(args, i, &val, "--max-buckets", (char*)NULL)) { max_buckets = (int)strict_strtol(val.c_str(), 10, &err); if (!err.empty()) { cerr << "ERROR: failed to parse max buckets: " << err << std::endl; return EINVAL; } max_buckets_specified = true; } else if (ceph_argparse_witharg(args, i, &val, "--max-entries", (char*)NULL)) { max_entries = (int)strict_strtol(val.c_str(), 10, &err); max_entries_specified = true; if (!err.empty()) { cerr << "ERROR: failed to parse max entries: " << err << std::endl; return EINVAL; } } else if (ceph_argparse_witharg(args, i, &val, "--max-size", (char*)NULL)) { max_size = strict_iec_cast(val.c_str(), &err); if (!err.empty()) { cerr << "ERROR: failed to parse max size: " << err << std::endl; return EINVAL; } have_max_size = true; } else if (ceph_argparse_witharg(args, i, &val, "--max-objects", (char*)NULL)) { max_objects = (int64_t)strict_strtoll(val.c_str(), 10, &err); if (!err.empty()) { cerr << "ERROR: failed to parse max objects: " << err << std::endl; return EINVAL; } have_max_objects = true; } else if (ceph_argparse_witharg(args, i, &val, "--date", "--time", (char*)NULL)) { date = val; if (end_date.empty()) end_date = date; } else if (ceph_argparse_witharg(args, i, &val, "--start-date", "--start-time", (char*)NULL)) { start_date = val; } else if (ceph_argparse_witharg(args, i, &val, "--end-date", "--end-time", (char*)NULL)) { end_date = val; } else if (ceph_argparse_witharg(args, i, &val, "--num-shards", (char*)NULL)) { num_shards = (int)strict_strtol(val.c_str(), 10, &err); if (!err.empty()) { cerr << "ERROR: failed to parse num shards: " << err << std::endl; return EINVAL; } num_shards_specified = true; } else if (ceph_argparse_witharg(args, i, &val, "--max-concurrent-ios", (char*)NULL)) { max_concurrent_ios = (int)strict_strtol(val.c_str(), 10, &err); if (!err.empty()) { cerr << "ERROR: failed to parse max concurrent ios: " << err << std::endl; return EINVAL; } } else if (ceph_argparse_witharg(args, i, &val, "--orphan-stale-secs", (char*)NULL)) { orphan_stale_secs = (uint64_t)strict_strtoll(val.c_str(), 10, &err); if (!err.empty()) { cerr << "ERROR: failed to parse orphan stale secs: " << err << std::endl; return EINVAL; } } else if (ceph_argparse_witharg(args, i, &val, "--shard-id", (char*)NULL)) { shard_id = (int)strict_strtol(val.c_str(), 10, &err); if (!err.empty()) { cerr << "ERROR: failed to parse shard id: " << err << std::endl; return EINVAL; } specified_shard_id = true; } else if (ceph_argparse_witharg(args, i, &val, "--daemon-id", (char*)NULL)) { daemon_id = val; specified_daemon_id = true; } else if (ceph_argparse_witharg(args, i, &val, "--access", (char*)NULL)) { access = val; perm_mask = rgw_str_to_perm(access.c_str()); set_perm = true; } else if (ceph_argparse_witharg(args, i, &val, "--temp-url-key", (char*)NULL)) { temp_url_keys[0] = val; set_temp_url_key = true; } else if (ceph_argparse_witharg(args, i, &val, "--temp-url-key2", "--temp-url-key-2", (char*)NULL)) { temp_url_keys[1] = val; set_temp_url_key = true; } else if (ceph_argparse_witharg(args, i, &val, "--bucket-id", (char*)NULL)) { bucket_id = val; if (bucket_id.empty()) { cerr << "bad bucket-id" << std::endl; usage(); assert(false); } } else if (ceph_argparse_witharg(args, i, &val, "--format", (char*)NULL)) { format = val; } else if (ceph_argparse_witharg(args, i, &val, "--categories", (char*)NULL)) { string cat_str = val; list cat_list; list::iterator iter; get_str_list(cat_str, cat_list); for (iter = cat_list.begin(); iter != cat_list.end(); ++iter) { categories[*iter] = true; } } else if (ceph_argparse_binary_flag(args, i, &delete_child_objects, NULL, "--purge-objects", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &pretty_format, NULL, "--pretty-format", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &purge_data, NULL, "--purge-data", (char*)NULL)) { delete_child_objects = purge_data; } else if (ceph_argparse_binary_flag(args, i, &purge_keys, NULL, "--purge-keys", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &yes_i_really_mean_it, NULL, "--yes-i-really-mean-it", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &fix, NULL, "--fix", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &remove_bad, NULL, "--remove-bad", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &check_head_obj_locator, NULL, "--check-head-obj-locator", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &check_objects, NULL, "--check-objects", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &sync_stats, NULL, "--sync-stats", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &reset_stats, NULL, "--reset-stats", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &include_all, NULL, "--include-all", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &extra_info, NULL, "--extra-info", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &bypass_gc, NULL, "--bypass-gc", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &warnings_only, NULL, "--warnings-only", (char*)NULL)) { // do nothing } else if (ceph_argparse_binary_flag(args, i, &inconsistent_index, NULL, "--inconsistent-index", (char*)NULL)) { // do nothing } else if (ceph_argparse_witharg(args, i, &val, "--caps", (char*)NULL)) { caps = val; } else if (ceph_argparse_witharg(args, i, &val, "-i", "--infile", (char*)NULL)) { infile = val; } else if (ceph_argparse_witharg(args, i, &val, "--metadata-key", (char*)NULL)) { metadata_key = val; } else if (ceph_argparse_witharg(args, i, &val, "--marker", (char*)NULL)) { marker = val; } else if (ceph_argparse_witharg(args, i, &val, "--start-marker", (char*)NULL)) { start_marker = val; } else if (ceph_argparse_witharg(args, i, &val, "--end-marker", (char*)NULL)) { end_marker = val; } else if (ceph_argparse_witharg(args, i, &val, "--quota-scope", (char*)NULL)) { quota_scope = val; } else if (ceph_argparse_witharg(args, i, &val, "--replica-log-type", (char*)NULL)) { replica_log_type_str = val; replica_log_type = get_replicalog_type(replica_log_type_str); if (replica_log_type == ReplicaLog_Invalid) { cerr << "ERROR: invalid replica log type" << std::endl; return EINVAL; } } else if (ceph_argparse_witharg(args, i, &val, "--index-type", (char*)NULL)) { string index_type_str = val; bi_index_type = get_bi_index_type(index_type_str); if (bi_index_type == InvalidIdx) { cerr << "ERROR: invalid bucket index entry type" << std::endl; return EINVAL; } } else if (ceph_argparse_binary_flag(args, i, &is_master_int, NULL, "--master", (char*)NULL)) { is_master = (bool)is_master_int; is_master_set = true; } else if (ceph_argparse_binary_flag(args, i, &set_default, NULL, "--default", (char*)NULL)) { /* do nothing */ } else if (ceph_argparse_binary_flag(args, i, &read_only_int, NULL, "--read-only", (char*)NULL)) { read_only = (bool)read_only_int; is_read_only_set = true; } else if (ceph_argparse_witharg(args, i, &val, "--master-url", (char*)NULL)) { master_url = val; } else if (ceph_argparse_witharg(args, i, &val, "--master-zonegroup", (char*)NULL)) { master_zonegroup = val; } else if (ceph_argparse_witharg(args, i, &val, "--master-zone", (char*)NULL)) { master_zone = val; } else if (ceph_argparse_witharg(args, i, &val, "--period", (char*)NULL)) { period_id = val; } else if (ceph_argparse_witharg(args, i, &val, "--epoch", (char*)NULL)) { period_epoch = val; } else if (ceph_argparse_witharg(args, i, &val, "--remote", (char*)NULL)) { remote = val; } else if (ceph_argparse_witharg(args, i, &val, "--url", (char*)NULL)) { url = val; } else if (ceph_argparse_witharg(args, i, &val, "--realm-id", (char*)NULL)) { realm_id = val; } else if (ceph_argparse_witharg(args, i, &val, "--realm-new-name", (char*)NULL)) { realm_new_name = val; } else if (ceph_argparse_witharg(args, i, &val, "--zonegroup-id", (char*)NULL)) { zonegroup_id = val; } else if (ceph_argparse_witharg(args, i, &val, "--zonegroup-new-name", (char*)NULL)) { zonegroup_new_name = val; } else if (ceph_argparse_witharg(args, i, &val, "--placement-id", (char*)NULL)) { placement_id = val; } else if (ceph_argparse_witharg(args, i, &val, "--tags", (char*)NULL)) { get_str_list(val, tags); } else if (ceph_argparse_witharg(args, i, &val, "--tags-add", (char*)NULL)) { get_str_list(val, tags_add); } else if (ceph_argparse_witharg(args, i, &val, "--tags-rm", (char*)NULL)) { get_str_list(val, tags_rm); } else if (ceph_argparse_witharg(args, i, &val, "--api-name", (char*)NULL)) { api_name = val; } else if (ceph_argparse_witharg(args, i, &val, "--zone-id", (char*)NULL)) { zone_id = val; } else if (ceph_argparse_witharg(args, i, &val, "--zone-new-name", (char*)NULL)) { zone_new_name = val; } else if (ceph_argparse_witharg(args, i, &val, "--endpoints", (char*)NULL)) { get_str_list(val, endpoints); } else if (ceph_argparse_witharg(args, i, &val, "--sync-from", (char*)NULL)) { get_str_list(val, sync_from); } else if (ceph_argparse_witharg(args, i, &val, "--sync-from-rm", (char*)NULL)) { get_str_list(val, sync_from_rm); } else if (ceph_argparse_binary_flag(args, i, &tmp_int, NULL, "--sync-from-all", (char*)NULL)) { sync_from_all = (bool)tmp_int; sync_from_all_specified = true; } else if (ceph_argparse_witharg(args, i, &val, "--source-zone", (char*)NULL)) { source_zone_name = val; } else if (ceph_argparse_witharg(args, i, &val, "--tier-type", (char*)NULL)) { tier_type = val; tier_type_specified = true; } else if (ceph_argparse_witharg(args, i, &val, "--tier-config", (char*)NULL)) { parse_tier_config_param(val, tier_config_add); } else if (ceph_argparse_witharg(args, i, &val, "--tier-config-rm", (char*)NULL)) { parse_tier_config_param(val, tier_config_rm); } else if (ceph_argparse_witharg(args, i, &val, "--index-pool", (char*)NULL)) { index_pool = val; } else if (ceph_argparse_witharg(args, i, &val, "--data-pool", (char*)NULL)) { data_pool = val; } else if (ceph_argparse_witharg(args, i, &val, "--data-extra-pool", (char*)NULL)) { data_extra_pool = val; } else if (ceph_argparse_witharg(args, i, &val, "--placement-index-type", (char*)NULL)) { if (val == "normal") { placement_index_type = RGWBIType_Normal; } else if (val == "indexless") { placement_index_type = RGWBIType_Indexless; } else { placement_index_type = (RGWBucketIndexType)strict_strtol(val.c_str(), 10, &err); if (!err.empty()) { cerr << "ERROR: failed to parse index type index: " << err << std::endl; return EINVAL; } } index_type_specified = true; } else if (ceph_argparse_witharg(args, i, &val, "--compression", (char*)NULL)) { compression_type = val; } else if (ceph_argparse_witharg(args, i, &val, "--role-name", (char*)NULL)) { role_name = val; } else if (ceph_argparse_witharg(args, i, &val, "--path", (char*)NULL)) { path = val; } else if (ceph_argparse_witharg(args, i, &val, "--assume-role-policy-doc", (char*)NULL)) { assume_role_doc = val; } else if (ceph_argparse_witharg(args, i, &val, "--policy-name", (char*)NULL)) { policy_name = val; } else if (ceph_argparse_witharg(args, i, &val, "--policy-doc", (char*)NULL)) { perm_policy_doc = val; } else if (ceph_argparse_witharg(args, i, &val, "--path-prefix", (char*)NULL)) { path_prefix = val; } else if (ceph_argparse_witharg(args, i, &val, "--trim-delay-ms", (char*)NULL)) { trim_delay_ms = atoi(val.c_str()); } else if (strncmp(*i, "-", 1) == 0) { cerr << "ERROR: invalid flag " << *i << std::endl; return EINVAL; } else { ++i; } } if (args.empty()) { usage(); assert(false); } else { const char *prev_cmd = NULL; const char *prev_prev_cmd = NULL; std::vector::iterator i ; for (i = args.begin(); i != args.end(); ++i) { opt_cmd = get_cmd(*i, prev_cmd, prev_prev_cmd, &need_more); if (opt_cmd < 0) { cerr << "unrecognized arg " << *i << std::endl; usage(); assert(false); } if (!need_more) { ++i; break; } prev_prev_cmd = prev_cmd; prev_cmd = *i; } if (opt_cmd == OPT_NO_CMD) { usage(); assert(false); } /* some commands may have an optional extra param */ if (i != args.end()) { switch (opt_cmd) { case OPT_METADATA_GET: case OPT_METADATA_PUT: case OPT_METADATA_RM: case OPT_METADATA_LIST: metadata_key = *i; break; default: break; } } if (tenant.empty()) { tenant = user_id.tenant; } else { if (user_id.empty() && opt_cmd != OPT_ROLE_CREATE && opt_cmd != OPT_ROLE_DELETE && opt_cmd != OPT_ROLE_GET && opt_cmd != OPT_ROLE_MODIFY && opt_cmd != OPT_ROLE_LIST && opt_cmd != OPT_ROLE_POLICY_PUT && opt_cmd != OPT_ROLE_POLICY_LIST && opt_cmd != OPT_ROLE_POLICY_GET && opt_cmd != OPT_ROLE_POLICY_DELETE && opt_cmd != OPT_RESHARD_ADD && opt_cmd != OPT_RESHARD_CANCEL && opt_cmd != OPT_RESHARD_STATUS) { cerr << "ERROR: --tenant is set, but there's no user ID" << std::endl; return EINVAL; } user_id.tenant = tenant; } /* check key parameter conflict */ if ((!access_key.empty()) && gen_access_key) { cerr << "ERROR: key parameter conflict, --access-key & --gen-access-key" << std::endl; return EINVAL; } if ((!secret_key.empty()) && gen_secret_key) { cerr << "ERROR: key parameter conflict, --secret & --gen-secret" << std::endl; return EINVAL; } } // default to pretty json if (format.empty()) { format = "json"; pretty_format = true; } if (format == "xml") formatter = new XMLFormatter(pretty_format); else if (format == "json") formatter = new JSONFormatter(pretty_format); else { cerr << "unrecognized format: " << format << std::endl; usage(); assert(false); } realm_name = g_conf->rgw_realm; zone_name = g_conf->rgw_zone; zonegroup_name = g_conf->rgw_zonegroup; RGWStreamFlusher f(formatter, cout); // not a raw op if 'period update' needs to commit to master bool raw_period_update = opt_cmd == OPT_PERIOD_UPDATE && !commit; std::set raw_storage_ops_list = {OPT_ZONEGROUP_ADD, OPT_ZONEGROUP_CREATE, OPT_ZONEGROUP_DELETE, OPT_ZONEGROUP_GET, OPT_ZONEGROUP_LIST, OPT_ZONEGROUP_SET, OPT_ZONEGROUP_DEFAULT, OPT_ZONEGROUP_RENAME, OPT_ZONEGROUP_MODIFY, OPT_ZONEGROUP_REMOVE, OPT_ZONEGROUP_PLACEMENT_ADD, OPT_ZONEGROUP_PLACEMENT_RM, OPT_ZONEGROUP_PLACEMENT_MODIFY, OPT_ZONEGROUP_PLACEMENT_LIST, OPT_ZONEGROUP_PLACEMENT_DEFAULT, OPT_ZONE_CREATE, OPT_ZONE_DELETE, OPT_ZONE_GET, OPT_ZONE_SET, OPT_ZONE_RENAME, OPT_ZONE_LIST, OPT_ZONE_MODIFY, OPT_ZONE_DEFAULT, OPT_ZONE_PLACEMENT_ADD, OPT_ZONE_PLACEMENT_RM, OPT_ZONE_PLACEMENT_MODIFY, OPT_ZONE_PLACEMENT_LIST, OPT_REALM_CREATE, OPT_PERIOD_DELETE, OPT_PERIOD_GET, OPT_PERIOD_PULL, OPT_PERIOD_GET_CURRENT, OPT_PERIOD_LIST, OPT_GLOBAL_QUOTA_GET, OPT_GLOBAL_QUOTA_SET, OPT_GLOBAL_QUOTA_ENABLE, OPT_GLOBAL_QUOTA_DISABLE, OPT_REALM_DELETE, OPT_REALM_GET, OPT_REALM_LIST, OPT_REALM_LIST_PERIODS, OPT_REALM_GET_DEFAULT, OPT_REALM_REMOVE, OPT_REALM_RENAME, OPT_REALM_SET, OPT_REALM_DEFAULT, OPT_REALM_PULL}; std::set readonly_ops_list = { OPT_USER_INFO, OPT_USER_STATS, OPT_BUCKETS_LIST, OPT_BUCKET_LIMIT_CHECK, OPT_BUCKET_STATS, OPT_BUCKET_SYNC_STATUS, OPT_LOG_LIST, OPT_LOG_SHOW, OPT_USAGE_SHOW, OPT_OBJECT_STAT, OPT_BI_GET, OPT_BI_LIST, OPT_OLH_GET, OPT_OLH_READLOG, OPT_GC_LIST, OPT_LC_LIST, OPT_ORPHANS_LIST_JOBS, OPT_ZONEGROUP_GET, OPT_ZONEGROUP_LIST, OPT_ZONEGROUP_PLACEMENT_LIST, OPT_ZONE_GET, OPT_ZONE_LIST, OPT_ZONE_PLACEMENT_LIST, OPT_METADATA_GET, OPT_METADATA_LIST, OPT_METADATA_SYNC_STATUS, OPT_MDLOG_LIST, OPT_MDLOG_STATUS, OPT_SYNC_ERROR_LIST, OPT_BILOG_LIST, OPT_BILOG_STATUS, OPT_DATA_SYNC_STATUS, OPT_DATALOG_LIST, OPT_DATALOG_STATUS, OPT_OPSTATE_LIST, OPT_REPLICALOG_GET, OPT_REALM_GET, OPT_REALM_GET_DEFAULT, OPT_REALM_LIST, OPT_REALM_LIST_PERIODS, OPT_PERIOD_GET, OPT_PERIOD_GET_CURRENT, OPT_PERIOD_LIST, OPT_GLOBAL_QUOTA_GET, OPT_SYNC_STATUS, OPT_ROLE_GET, OPT_ROLE_LIST, OPT_ROLE_POLICY_LIST, OPT_ROLE_POLICY_GET, OPT_RESHARD_LIST, OPT_RESHARD_STATUS, }; bool raw_storage_op = (raw_storage_ops_list.find(opt_cmd) != raw_storage_ops_list.end() || raw_period_update); bool need_cache = readonly_ops_list.find(opt_cmd) == readonly_ops_list.end(); if (raw_storage_op) { store = RGWStoreManager::get_raw_storage(g_ceph_context); } else { store = RGWStoreManager::get_storage(g_ceph_context, false, false, false, false, false, need_cache && g_conf->rgw_cache_enabled); } if (!store) { cerr << "couldn't init storage provider" << std::endl; return 5; //EIO } if (!source_zone_name.empty()) { if (!store->find_zone_id_by_name(source_zone_name, &source_zone)) { cerr << "WARNING: cannot find source zone id for name=" << source_zone_name << std::endl; source_zone = source_zone_name; } } rgw_user_init(store); rgw_bucket_init(store->meta_mgr); struct rgw_curl_setup { rgw_curl_setup() { rgw::curl::setup_curl(boost::none); } ~rgw_curl_setup() { rgw::curl::cleanup_curl(); } } curl_cleanup; StoreDestructor store_destructor(store); if (raw_storage_op) { switch (opt_cmd) { case OPT_PERIOD_DELETE: { if (period_id.empty()) { cerr << "missing period id" << std::endl; return EINVAL; } RGWPeriod period(period_id); int ret = period.init(g_ceph_context, store); if (ret < 0) { cerr << "period.init failed: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = period.delete_obj(); if (ret < 0) { cerr << "ERROR: couldn't delete period: " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_PERIOD_GET: { epoch_t epoch = 0; if (!period_epoch.empty()) { epoch = atoi(period_epoch.c_str()); } if (staging) { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0 ) { cerr << "Error initializing realm " << cpp_strerror(-ret) << std::endl; return -ret; } realm_id = realm.get_id(); realm_name = realm.get_name(); period_id = RGWPeriod::get_staging_id(realm_id); epoch = 1; } RGWPeriod period(period_id, epoch); int ret = period.init(g_ceph_context, store, realm_id, realm_name); if (ret < 0) { cerr << "period init failed: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("period", period, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_PERIOD_GET_CURRENT: { int ret = read_current_period_id(store, realm_id, realm_name, &period_id); if (ret < 0) { return -ret; } formatter->open_object_section("period_get_current"); encode_json("current_period", period_id, formatter); formatter->close_section(); formatter->flush(cout); } break; case OPT_PERIOD_LIST: { list periods; int ret = store->list_periods(periods); if (ret < 0) { cerr << "failed to list periods: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_object_section("periods_list"); encode_json("periods", periods, formatter); formatter->close_section(); formatter->flush(cout); cout << std::endl; } break; case OPT_PERIOD_UPDATE: { int ret = update_period(realm_id, realm_name, period_id, period_epoch, commit, remote, url, access_key, secret_key, formatter, yes_i_really_mean_it); if (ret < 0) { return -ret; } } break; case OPT_PERIOD_PULL: { boost::optional conn; RGWRESTConn *remote_conn = nullptr; if (url.empty()) { // load current period for endpoints RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init realm: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWPeriod current_period(realm.get_current_period()); ret = current_period.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init current period: " << cpp_strerror(-ret) << std::endl; return -ret; } if (remote.empty()) { // use realm master zone as remote remote = current_period.get_master_zone(); } conn = get_remote_conn(store, current_period.get_map(), remote); if (!conn) { cerr << "failed to find a zone or zonegroup for remote " << remote << std::endl; return -ENOENT; } remote_conn = &*conn; } RGWPeriod period; int ret = do_period_pull(remote_conn, url, access_key, secret_key, realm_id, realm_name, period_id, period_epoch, &period); if (ret < 0) { cerr << "period pull failed: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("period", period, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_GLOBAL_QUOTA_GET: case OPT_GLOBAL_QUOTA_SET: case OPT_GLOBAL_QUOTA_ENABLE: case OPT_GLOBAL_QUOTA_DISABLE: { if (realm_id.empty()) { RGWRealm realm(g_ceph_context, store); if (!realm_name.empty()) { // look up realm_id for the given realm_name int ret = realm.read_id(realm_name, realm_id); if (ret < 0) { cerr << "ERROR: failed to read realm for " << realm_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } } else { // use default realm_id when none is given int ret = realm.read_default_id(realm_id); if (ret < 0 && ret != -ENOENT) { // on ENOENT, use empty realm_id cerr << "ERROR: failed to read default realm: " << cpp_strerror(-ret) << std::endl; return -ret; } } } RGWPeriodConfig period_config; int ret = period_config.read(store, realm_id); if (ret < 0 && ret != -ENOENT) { cerr << "ERROR: failed to read period config: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_object_section("period_config"); if (quota_scope == "bucket") { set_quota_info(period_config.bucket_quota, opt_cmd, max_size, max_objects, have_max_size, have_max_objects); encode_json("bucket quota", period_config.bucket_quota, formatter); } else if (quota_scope == "user") { set_quota_info(period_config.user_quota, opt_cmd, max_size, max_objects, have_max_size, have_max_objects); encode_json("user quota", period_config.user_quota, formatter); } else if (quota_scope.empty() && opt_cmd == OPT_GLOBAL_QUOTA_GET) { // if no scope is given for GET, print both encode_json("bucket quota", period_config.bucket_quota, formatter); encode_json("user quota", period_config.user_quota, formatter); } else { cerr << "ERROR: invalid quota scope specification. Please specify " "either --quota-scope=bucket, or --quota-scope=user" << std::endl; return EINVAL; } formatter->close_section(); if (opt_cmd != OPT_GLOBAL_QUOTA_GET) { // write the modified period config ret = period_config.write(store, realm_id); if (ret < 0) { cerr << "ERROR: failed to write period config: " << cpp_strerror(-ret) << std::endl; return -ret; } if (!realm_id.empty()) { cout << "Global quota changes saved. Use 'period update' to apply " "them to the staging period, and 'period commit' to commit the " "new period." << std::endl; } else { cout << "Global quota changes saved. They will take effect as " "the gateways are restarted." << std::endl; } } formatter->flush(cout); cout << std::endl; } break; case OPT_REALM_CREATE: { if (realm_name.empty()) { cerr << "missing realm name" << std::endl; return EINVAL; } RGWRealm realm(realm_name, g_ceph_context, store); int ret = realm.create(); if (ret < 0) { cerr << "ERROR: couldn't create realm " << realm_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } if (set_default) { ret = realm.set_as_default(); if (ret < 0) { cerr << "failed to set realm " << realm_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("realm", realm, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_REALM_DELETE: { RGWRealm realm(realm_id, realm_name); if (realm_name.empty() && realm_id.empty()) { cerr << "missing realm name or id" << std::endl; return EINVAL; } int ret = realm.init(g_ceph_context, store); if (ret < 0) { cerr << "realm.init failed: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = realm.delete_obj(); if (ret < 0) { cerr << "ERROR: couldn't : " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_REALM_GET: { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0) { if (ret == -ENOENT && realm_name.empty() && realm_id.empty()) { cerr << "missing realm name or id, or default realm not found" << std::endl; } else { cerr << "realm.init failed: " << cpp_strerror(-ret) << std::endl; } return -ret; } encode_json("realm", realm, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_REALM_GET_DEFAULT: { RGWRealm realm(g_ceph_context, store); string default_id; int ret = realm.read_default_id(default_id); if (ret == -ENOENT) { cout << "No default realm is set" << std::endl; return -ret; } else if (ret < 0) { cerr << "Error reading default realm:" << cpp_strerror(-ret) << std::endl; return -ret; } cout << "default realm: " << default_id << std::endl; } break; case OPT_REALM_LIST: { RGWRealm realm(g_ceph_context, store); string default_id; int ret = realm.read_default_id(default_id); if (ret < 0 && ret != -ENOENT) { cerr << "could not determine default realm: " << cpp_strerror(-ret) << std::endl; } list realms; ret = store->list_realms(realms); if (ret < 0) { cerr << "failed to list realms: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_object_section("realms_list"); encode_json("default_info", default_id, formatter); encode_json("realms", realms, formatter); formatter->close_section(); formatter->flush(cout); cout << std::endl; } break; case OPT_REALM_LIST_PERIODS: { int ret = read_current_period_id(store, realm_id, realm_name, &period_id); if (ret < 0) { return -ret; } list periods; ret = store->list_periods(period_id, periods); if (ret < 0) { cerr << "list periods failed: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_object_section("realm_periods_list"); encode_json("current_period", period_id, formatter); encode_json("periods", periods, formatter); formatter->close_section(); formatter->flush(cout); cout << std::endl; } break; case OPT_REALM_RENAME: { RGWRealm realm(realm_id, realm_name); if (realm_new_name.empty()) { cerr << "missing realm new name" << std::endl; return EINVAL; } if (realm_name.empty() && realm_id.empty()) { cerr << "missing realm name or id" << std::endl; return EINVAL; } int ret = realm.init(g_ceph_context, store); if (ret < 0) { cerr << "realm.init failed: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = realm.rename(realm_new_name); if (ret < 0) { cerr << "realm.rename failed: " << cpp_strerror(-ret) << std::endl; return -ret; } cout << "Realm name updated. Note that this change only applies to " "the current cluster, so this command must be run separately " "on each of the realm's other clusters." << std::endl; } break; case OPT_REALM_SET: { if (realm_id.empty() && realm_name.empty()) { cerr << "no realm name or id provided" << std::endl; return EINVAL; } RGWRealm realm(realm_id, realm_name); bool new_realm = false; int ret = realm.init(g_ceph_context, store); if (ret < 0 && ret != -ENOENT) { cerr << "failed to init realm: " << cpp_strerror(-ret) << std::endl; return -ret; } else if (ret == -ENOENT) { new_realm = true; } ret = read_decode_json(infile, realm); if (ret < 0) { return 1; } if (!realm_name.empty() && realm.get_name() != realm_name) { cerr << "mismatch between --rgw-realm " << realm_name << " and json input file name " << realm.get_name() << std::endl; return EINVAL; } /* new realm */ if (new_realm) { cout << "clearing period and epoch for new realm" << std::endl; realm.clear_current_period_and_epoch(); ret = realm.create(); if (ret < 0) { cerr << "ERROR: couldn't create new realm: " << cpp_strerror(-ret) << std::endl; return 1; } } else { ret = realm.update(); if (ret < 0) { cerr << "ERROR: couldn't store realm info: " << cpp_strerror(-ret) << std::endl; return 1; } } if (set_default) { ret = realm.set_as_default(); if (ret < 0) { cerr << "failed to set realm " << realm_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("realm", realm, formatter); formatter->flush(cout); } break; case OPT_REALM_DEFAULT: { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init realm: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = realm.set_as_default(); if (ret < 0) { cerr << "failed to set realm as default: " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_REALM_PULL: { if (url.empty()) { cerr << "A --url must be provided." << std::endl; return EINVAL; } RGWEnv env; req_info info(g_ceph_context, &env); info.method = "GET"; info.request_uri = "/admin/realm"; map ¶ms = info.args.get_params(); if (!realm_id.empty()) params["id"] = realm_id; if (!realm_name.empty()) params["name"] = realm_name; bufferlist bl; JSONParser p; int ret = send_to_url(url, access_key, secret_key, info, bl, p); if (ret < 0) { cerr << "request failed: " << cpp_strerror(-ret) << std::endl; if (ret == -EACCES) { cerr << "If the realm has been changed on the master zone, the " "master zone's gateway may need to be restarted to recognize " "this user." << std::endl; } return -ret; } RGWRealm realm; realm.init(g_ceph_context, store, false); try { decode_json_obj(realm, &p); } catch (JSONDecoder::err& e) { cerr << "failed to decode JSON response: " << e.message << std::endl; return EINVAL; } RGWPeriod period; auto& current_period = realm.get_current_period(); if (!current_period.empty()) { // pull the latest epoch of the realm's current period ret = do_period_pull(nullptr, url, access_key, secret_key, realm_id, realm_name, current_period, "", &period); if (ret < 0) { cerr << "could not fetch period " << current_period << std::endl; return -ret; } } ret = realm.create(false); if (ret < 0 && ret != -EEXIST) { cerr << "Error storing realm " << realm.get_id() << ": " << cpp_strerror(ret) << std::endl; return -ret; } else if (ret ==-EEXIST) { ret = realm.update(); if (ret < 0) { cerr << "Error storing realm " << realm.get_id() << ": " << cpp_strerror(ret) << std::endl; } } if (set_default) { ret = realm.set_as_default(); if (ret < 0) { cerr << "failed to set realm " << realm_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("realm", realm, formatter); formatter->flush(cout); cout << std::endl; } return 0; case OPT_ZONEGROUP_ADD: { if (zonegroup_id.empty() && zonegroup_name.empty()) { cerr << "no zonegroup name or id provided" << std::endl; return EINVAL; } RGWZoneGroup zonegroup(zonegroup_id,zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to initialize zonegroup " << zonegroup_name << " id " << zonegroup_id << " :" << cpp_strerror(-ret) << std::endl; return -ret; } RGWZoneParams zone(zone_id, zone_name); ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "unable to initialize zone: " << cpp_strerror(-ret) << std::endl; return -ret; } if (zone.realm_id != zonegroup.realm_id) { zone.realm_id = zonegroup.realm_id; ret = zone.update(); if (ret < 0) { cerr << "failed to save zone info: " << cpp_strerror(-ret) << std::endl; return -ret; } } string *ptier_type = (tier_type_specified ? &tier_type : nullptr); zone.tier_config = tier_config_add; bool *psync_from_all = (sync_from_all_specified ? &sync_from_all : nullptr); ret = zonegroup.add_zone(zone, (is_master_set ? &is_master : NULL), (is_read_only_set ? &read_only : NULL), endpoints, ptier_type, psync_from_all, sync_from, sync_from_rm); if (ret < 0) { cerr << "failed to add zone " << zone_name << " to zonegroup " << zonegroup.get_name() << ": " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("zonegroup", zonegroup, formatter); formatter->flush(cout); } break; case OPT_ZONEGROUP_CREATE: { if (zonegroup_name.empty()) { cerr << "Missing zonegroup name" << std::endl; return EINVAL; } RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init realm: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWZoneGroup zonegroup(zonegroup_name, is_master, g_ceph_context, store, realm.get_id(), endpoints); zonegroup.api_name = (api_name.empty() ? zonegroup_name : api_name); ret = zonegroup.create(); if (ret < 0) { cerr << "failed to create zonegroup " << zonegroup_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } if (set_default) { ret = zonegroup.set_as_default(); if (ret < 0) { cerr << "failed to set zonegroup " << zonegroup_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("zonegroup", zonegroup, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_ZONEGROUP_DEFAULT: { if (zonegroup_id.empty() && zonegroup_name.empty()) { cerr << "no zonegroup name or id provided" << std::endl; return EINVAL; } RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = zonegroup.set_as_default(); if (ret < 0) { cerr << "failed to set zonegroup as default: " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_ZONEGROUP_DELETE: { if (zonegroup_id.empty() && zonegroup_name.empty()) { cerr << "no zonegroup name or id provided" << std::endl; return EINVAL; } RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = zonegroup.delete_obj(); if (ret < 0) { cerr << "ERROR: couldn't delete zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_ZONEGROUP_GET: { RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("zonegroup", zonegroup, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_ZONEGROUP_LIST: { RGWZoneGroup zonegroup; int ret = zonegroup.init(g_ceph_context, store, false); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } list zonegroups; ret = store->list_zonegroups(zonegroups); if (ret < 0) { cerr << "failed to list zonegroups: " << cpp_strerror(-ret) << std::endl; return -ret; } string default_zonegroup; ret = zonegroup.read_default_id(default_zonegroup); if (ret < 0 && ret != -ENOENT) { cerr << "could not determine default zonegroup: " << cpp_strerror(-ret) << std::endl; } formatter->open_object_section("zonegroups_list"); encode_json("default_info", default_zonegroup, formatter); encode_json("zonegroups", zonegroups, formatter); formatter->close_section(); formatter->flush(cout); cout << std::endl; } break; case OPT_ZONEGROUP_MODIFY: { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init realm: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } bool need_update = false; if (!master_zone.empty()) { zonegroup.master_zone = master_zone; need_update = true; } if (is_master_set) { zonegroup.update_master(is_master); need_update = true; } if (!endpoints.empty()) { zonegroup.endpoints = endpoints; need_update = true; } if (!api_name.empty()) { zonegroup.api_name = api_name; need_update = true; } if (!realm_id.empty()) { zonegroup.realm_id = realm_id; need_update = true; } else if (!realm_name.empty()) { // get realm id from name RGWRealm realm{g_ceph_context, store}; ret = realm.read_id(realm_name, zonegroup.realm_id); if (ret < 0) { cerr << "failed to find realm by name " << realm_name << std::endl; return -ret; } need_update = true; } if (need_update) { ret = zonegroup.update(); if (ret < 0) { cerr << "failed to update zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } } if (set_default) { ret = zonegroup.set_as_default(); if (ret < 0) { cerr << "failed to set zonegroup " << zonegroup_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("zonegroup", zonegroup, formatter); formatter->flush(cout); } break; case OPT_ZONEGROUP_SET: { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); bool default_realm_not_exist = (ret == -ENOENT && realm_id.empty() && realm_name.empty()); if (ret < 0 && !default_realm_not_exist ) { cerr << "failed to init realm: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWZoneGroup zonegroup; ret = zonegroup.init(g_ceph_context, store, false); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = read_decode_json(infile, zonegroup); if (ret < 0) { return 1; } if (zonegroup.realm_id.empty() && !default_realm_not_exist) { zonegroup.realm_id = realm.get_id(); } ret = zonegroup.create(); if (ret < 0 && ret != -EEXIST) { cerr << "ERROR: couldn't create zonegroup info: " << cpp_strerror(-ret) << std::endl; return 1; } else if (ret == -EEXIST) { ret = zonegroup.update(); if (ret < 0) { cerr << "ERROR: couldn't store zonegroup info: " << cpp_strerror(-ret) << std::endl; return 1; } } if (set_default) { ret = zonegroup.set_as_default(); if (ret < 0) { cerr << "failed to set zonegroup " << zonegroup_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("zonegroup", zonegroup, formatter); formatter->flush(cout); } break; case OPT_ZONEGROUP_REMOVE: { RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } if (zone_id.empty()) { if (zone_name.empty()) { cerr << "no --zone-id or --rgw-zone name provided" << std::endl; return EINVAL; } // look up zone id by name for (auto& z : zonegroup.zones) { if (zone_name == z.second.name) { zone_id = z.second.id; break; } } if (zone_id.empty()) { cerr << "zone name " << zone_name << " not found in zonegroup " << zonegroup.get_name() << std::endl; return ENOENT; } } ret = zonegroup.remove_zone(zone_id); if (ret < 0) { cerr << "failed to remove zone: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("zonegroup", zonegroup, formatter); formatter->flush(cout); } break; case OPT_ZONEGROUP_RENAME: { if (zonegroup_new_name.empty()) { cerr << " missing zonegroup new name" << std::endl; return EINVAL; } if (zonegroup_id.empty() && zonegroup_name.empty()) { cerr << "no zonegroup name or id provided" << std::endl; return EINVAL; } RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = zonegroup.rename(zonegroup_new_name); if (ret < 0) { cerr << "failed to rename zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_ZONEGROUP_PLACEMENT_LIST: { RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("placement_targets", zonegroup.placement_targets, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_ZONEGROUP_PLACEMENT_ADD: case OPT_ZONEGROUP_PLACEMENT_MODIFY: case OPT_ZONEGROUP_PLACEMENT_RM: case OPT_ZONEGROUP_PLACEMENT_DEFAULT: { if (placement_id.empty()) { cerr << "ERROR: --placement-id not specified" << std::endl; return EINVAL; } RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } if (opt_cmd == OPT_ZONEGROUP_PLACEMENT_ADD) { RGWZoneGroupPlacementTarget target; target.name = placement_id; for (auto& t : tags) { target.tags.insert(t); } zonegroup.placement_targets[placement_id] = target; } else if (opt_cmd == OPT_ZONEGROUP_PLACEMENT_MODIFY) { RGWZoneGroupPlacementTarget& target = zonegroup.placement_targets[placement_id]; if (!tags.empty()) { target.tags.clear(); for (auto& t : tags) { target.tags.insert(t); } } target.name = placement_id; for (auto& t : tags_rm) { target.tags.erase(t); } for (auto& t : tags_add) { target.tags.insert(t); } } else if (opt_cmd == OPT_ZONEGROUP_PLACEMENT_RM) { zonegroup.placement_targets.erase(placement_id); } else if (opt_cmd == OPT_ZONEGROUP_PLACEMENT_DEFAULT) { if (!zonegroup.placement_targets.count(placement_id)) { cerr << "failed to find a zonegroup placement target named '" << placement_id << "'" << std::endl; return -ENOENT; } zonegroup.default_placement = placement_id; } zonegroup.post_process_params(); ret = zonegroup.update(); if (ret < 0) { cerr << "failed to update zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("placement_targets", zonegroup.placement_targets, formatter); formatter->flush(cout); } break; case OPT_ZONE_CREATE: { if (zone_name.empty()) { cerr << "zone name not provided" << std::endl; return EINVAL; } int ret; RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); /* if the user didn't provide zonegroup info , create stand alone zone */ if (!zonegroup_id.empty() || !zonegroup_name.empty()) { ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "unable to initialize zonegroup " << zonegroup_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } if (realm_id.empty() && realm_name.empty()) { realm_id = zonegroup.realm_id; } } RGWZoneParams zone(zone_id, zone_name); ret = zone.init(g_ceph_context, store, false); if (ret < 0) { cerr << "unable to initialize zone: " << cpp_strerror(-ret) << std::endl; return -ret; } zone.system_key.id = access_key; zone.system_key.key = secret_key; zone.realm_id = realm_id; zone.tier_config = tier_config_add; ret = zone.create(); if (ret < 0) { cerr << "failed to create zone " << zone_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } if (!zonegroup_id.empty() || !zonegroup_name.empty()) { string *ptier_type = (tier_type_specified ? &tier_type : nullptr); bool *psync_from_all = (sync_from_all_specified ? &sync_from_all : nullptr); ret = zonegroup.add_zone(zone, (is_master_set ? &is_master : NULL), (is_read_only_set ? &read_only : NULL), endpoints, ptier_type, psync_from_all, sync_from, sync_from_rm); if (ret < 0) { cerr << "failed to add zone " << zone_name << " to zonegroup " << zonegroup.get_name() << ": " << cpp_strerror(-ret) << std::endl; return -ret; } } if (set_default) { ret = zone.set_as_default(); if (ret < 0) { cerr << "failed to set zone " << zone_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("zone", zone, formatter); formatter->flush(cout); cout << std::endl; } break; case OPT_ZONE_DEFAULT: { RGWZoneGroup zonegroup(zonegroup_id,zonegroup_name); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "WARNING: failed to initialize zonegroup " << zonegroup_name << std::endl; } if (zone_id.empty() && zone_name.empty()) { cerr << "no zone name or id provided" << std::endl; return EINVAL; } RGWZoneParams zone(zone_id, zone_name); ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "unable to initialize zone: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = zone.set_as_default(); if (ret < 0) { cerr << "failed to set zone as default: " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_ZONE_DELETE: { if (zone_id.empty() && zone_name.empty()) { cerr << "no zone name or id provided" << std::endl; return EINVAL; } RGWZoneParams zone(zone_id, zone_name); int ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "unable to initialize zone: " << cpp_strerror(-ret) << std::endl; return -ret; } list zonegroups; ret = store->list_zonegroups(zonegroups); if (ret < 0) { cerr << "failed to list zonegroups: " << cpp_strerror(-ret) << std::endl; return -ret; } for (list::iterator iter = zonegroups.begin(); iter != zonegroups.end(); ++iter) { RGWZoneGroup zonegroup(string(), *iter); int ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "WARNING: failed to initialize zonegroup " << zonegroup_name << std::endl; continue; } ret = zonegroup.remove_zone(zone.get_id()); if (ret < 0 && ret != -ENOENT) { cerr << "failed to remove zone " << zone_name << " from zonegroup " << zonegroup.get_name() << ": " << cpp_strerror(-ret) << std::endl; } } ret = zone.delete_obj(); if (ret < 0) { cerr << "failed to delete zone " << zone_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } } break; case OPT_ZONE_GET: { RGWZoneParams zone(zone_id, zone_name); int ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "unable to initialize zone: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("zone", zone, formatter); formatter->flush(cout); } break; case OPT_ZONE_SET: { RGWZoneParams zone(zone_name); int ret = zone.init(g_ceph_context, store, false); if (ret < 0) { return -ret; } ret = zone.read(); if (ret < 0 && ret != -ENOENT) { cerr << "zone.read() returned ret=" << ret << std::endl; return -ret; } string orig_id = zone.get_id(); ret = read_decode_json(infile, zone); if (ret < 0) { return 1; } if(zone.realm_id.empty()) { RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0 && ret != -ENOENT) { cerr << "failed to init realm: " << cpp_strerror(-ret) << std::endl; return -ret; } zone.realm_id = realm.get_id(); } if( !zone_name.empty() && !zone.get_name().empty() && zone.get_name() != zone_name) { cerr << "Error: zone name" << zone_name << " is different than the zone name " << zone.get_name() << " in the provided json " << std::endl; return EINVAL; } if (zone.get_name().empty()) { zone.set_name(zone_name); if (zone.get_name().empty()) { cerr << "no zone name specified" << std::endl; return EINVAL; } } zone_name = zone.get_name(); if (zone.get_id().empty()) { zone.set_id(orig_id); } if (zone.get_id().empty()) { cerr << "no zone name id the json provided, assuming old format" << std::endl; if (zone_name.empty()) { cerr << "missing zone name" << std::endl; return EINVAL; } zone.set_name(zone_name); zone.set_id(zone_name); } cerr << "zone id " << zone.get_id(); ret = zone.fix_pool_names(); if (ret < 0) { cerr << "ERROR: couldn't fix zone: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = zone.write(false); if (ret < 0) { cerr << "ERROR: couldn't create zone: " << cpp_strerror(-ret) << std::endl; return 1; } if (set_default) { ret = zone.set_as_default(); if (ret < 0) { cerr << "failed to set zone " << zone_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("zone", zone, formatter); formatter->flush(cout); } break; case OPT_ZONE_LIST: { list zones; int ret = store->list_zones(zones); if (ret < 0) { cerr << "failed to list zones: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWZoneParams zone; ret = zone.init(g_ceph_context, store, false); if (ret < 0) { cerr << "failed to init zone: " << cpp_strerror(-ret) << std::endl; return -ret; } string default_zone; ret = zone.read_default_id(default_zone); if (ret < 0 && ret != -ENOENT) { cerr << "could not determine default zone: " << cpp_strerror(-ret) << std::endl; } formatter->open_object_section("zones_list"); encode_json("default_info", default_zone, formatter); encode_json("zones", zones, formatter); formatter->close_section(); formatter->flush(cout); cout << std::endl; } break; case OPT_ZONE_MODIFY: { RGWZoneParams zone(zone_id, zone_name); int ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zone: " << cpp_strerror(-ret) << std::endl; return -ret; } bool need_zone_update = false; if (!access_key.empty()) { zone.system_key.id = access_key; need_zone_update = true; } if (!secret_key.empty()) { zone.system_key.key = secret_key; need_zone_update = true; } if (!realm_id.empty()) { zone.realm_id = realm_id; need_zone_update = true; } else if (!realm_name.empty()) { // get realm id from name RGWRealm realm{g_ceph_context, store}; ret = realm.read_id(realm_name, zone.realm_id); if (ret < 0) { cerr << "failed to find realm by name " << realm_name << std::endl; return -ret; } need_zone_update = true; } for (auto add : tier_config_add) { zone.tier_config[add.first] = add.second; need_zone_update = true; } for (auto rm : tier_config_rm) { zone.tier_config.erase(rm.first); need_zone_update = true; } if (need_zone_update) { ret = zone.update(); if (ret < 0) { cerr << "failed to save zone info: " << cpp_strerror(-ret) << std::endl; return -ret; } } RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } string *ptier_type = (tier_type_specified ? &tier_type : nullptr); bool *psync_from_all = (sync_from_all_specified ? &sync_from_all : nullptr); ret = zonegroup.add_zone(zone, (is_master_set ? &is_master : NULL), (is_read_only_set ? &read_only : NULL), endpoints, ptier_type, psync_from_all, sync_from, sync_from_rm); if (ret < 0) { cerr << "failed to update zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = zonegroup.update(); if (ret < 0) { cerr << "failed to update zonegroup: " << cpp_strerror(-ret) << std::endl; return -ret; } if (set_default) { ret = zone.set_as_default(); if (ret < 0) { cerr << "failed to set zone " << zone_name << " as default: " << cpp_strerror(-ret) << std::endl; } } encode_json("zone", zone, formatter); formatter->flush(cout); } break; case OPT_ZONE_RENAME: { if (zone_new_name.empty()) { cerr << " missing zone new name" << std::endl; return EINVAL; } if (zone_id.empty() && zone_name.empty()) { cerr << "no zonegroup name or id provided" << std::endl; return EINVAL; } RGWZoneParams zone(zone_id,zone_name); int ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "unable to initialize zone: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = zone.rename(zone_new_name); if (ret < 0) { cerr << "failed to rename zone " << zone_name << " to " << zone_new_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } RGWZoneGroup zonegroup(zonegroup_id, zonegroup_name); ret = zonegroup.init(g_ceph_context, store); if (ret < 0) { cerr << "WARNING: failed to initialize zonegroup " << zonegroup_name << std::endl; } else { ret = zonegroup.rename_zone(zone); if (ret < 0 && ret ) { cerr << "Error in zonegroup rename for " << zone_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } } } break; case OPT_ZONE_PLACEMENT_ADD: case OPT_ZONE_PLACEMENT_MODIFY: case OPT_ZONE_PLACEMENT_RM: { if (placement_id.empty()) { cerr << "ERROR: --placement-id not specified" << std::endl; return EINVAL; } // validate compression type if (compression_type && *compression_type != "random" && !Compressor::get_comp_alg_type(*compression_type)) { std::cerr << "Unrecognized compression type" << std::endl; return EINVAL; } RGWZoneParams zone(zone_id, zone_name); int ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "failed to init zone: " << cpp_strerror(-ret) << std::endl; return -ret; } if (opt_cmd == OPT_ZONE_PLACEMENT_ADD) { // pool names are required if (!index_pool || index_pool->empty() || !data_pool || data_pool->empty()) { cerr << "ERROR: need to specify both --index-pool and --data-pool" << std::endl; return EINVAL; } RGWZonePlacementInfo& info = zone.placement_pools[placement_id]; info.index_pool = *index_pool; info.data_pool = *data_pool; if (data_extra_pool) { info.data_extra_pool = *data_extra_pool; } if (index_type_specified) { info.index_type = placement_index_type; } if (compression_type) { info.compression_type = *compression_type; } } else if (opt_cmd == OPT_ZONE_PLACEMENT_MODIFY) { auto p = zone.placement_pools.find(placement_id); if (p == zone.placement_pools.end()) { cerr << "ERROR: zone placement target '" << placement_id << "' not found" << std::endl; return -ENOENT; } auto& info = p->second; if (index_pool && !index_pool->empty()) { info.index_pool = *index_pool; } if (data_pool && !data_pool->empty()) { info.data_pool = *data_pool; } if (data_extra_pool) { info.data_extra_pool = *data_extra_pool; } if (index_type_specified) { info.index_type = placement_index_type; } if (compression_type) { info.compression_type = *compression_type; } } else if (opt_cmd == OPT_ZONE_PLACEMENT_RM) { zone.placement_pools.erase(placement_id); } ret = zone.update(); if (ret < 0) { cerr << "failed to save zone info: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("zone", zone, formatter); formatter->flush(cout); } break; case OPT_ZONE_PLACEMENT_LIST: { RGWZoneParams zone(zone_id, zone_name); int ret = zone.init(g_ceph_context, store); if (ret < 0) { cerr << "unable to initialize zone: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("placement_pools", zone.placement_pools, formatter); formatter->flush(cout); } break; } return 0; } if (!user_id.empty()) { user_op.set_user_id(user_id); bucket_op.set_user_id(user_id); } if (!display_name.empty()) user_op.set_display_name(display_name); if (!user_email.empty()) user_op.set_user_email(user_email); if (!access_key.empty()) user_op.set_access_key(access_key); if (!secret_key.empty()) user_op.set_secret_key(secret_key); if (!subuser.empty()) user_op.set_subuser(subuser); if (!caps.empty()) user_op.set_caps(caps); user_op.set_purge_data(purge_data); if (purge_keys) user_op.set_purge_keys(); if (gen_access_key) user_op.set_generate_key(); if (gen_secret_key) user_op.set_gen_secret(); // assume that a key pair should be created if (max_buckets_specified) user_op.set_max_buckets(max_buckets); if (admin_specified) user_op.set_admin(admin); if (system_specified) user_op.set_system(system); if (set_perm) user_op.set_perm(perm_mask); if (set_temp_url_key) { map::iterator iter = temp_url_keys.begin(); for (; iter != temp_url_keys.end(); ++iter) { user_op.set_temp_url_key(iter->second, iter->first); } } if (!op_mask_str.empty()) { uint32_t op_mask; int ret = rgw_parse_op_type_list(op_mask_str, &op_mask); if (ret < 0) { cerr << "failed to parse op_mask: " << cpp_strerror(-ret) << std::endl; return -ret; } user_op.set_op_mask(op_mask); } if (key_type != KEY_TYPE_UNDEFINED) user_op.set_key_type(key_type); // set suspension operation parameters if (opt_cmd == OPT_USER_ENABLE) user_op.set_suspension(false); else if (opt_cmd == OPT_USER_SUSPEND) user_op.set_suspension(true); // RGWUser to use for user operations RGWUser user; int ret = 0; if (!user_id.empty() || !subuser.empty()) { ret = user.init(store, user_op); if (ret < 0) { cerr << "user.init failed: " << cpp_strerror(-ret) << std::endl; return -ret; } } /* populate bucket operation */ bucket_op.set_bucket_name(bucket_name); bucket_op.set_object(object); bucket_op.set_check_objects(check_objects); bucket_op.set_delete_children(delete_child_objects); bucket_op.set_fix_index(fix); bucket_op.set_max_aio(max_concurrent_ios); // required to gather errors from operations std::string err_msg; bool output_user_info = true; switch (opt_cmd) { case OPT_USER_INFO: break; case OPT_USER_CREATE: if (!user_op.has_existing_user()) { user_op.set_generate_key(); // generate a new key by default } ret = user.add(user_op, &err_msg); if (ret < 0) { cerr << "could not create user: " << err_msg << std::endl; if (ret == -ERR_INVALID_TENANT_NAME) ret = -EINVAL; return -ret; } if (!subuser.empty()) { ret = user.subusers.add(user_op, &err_msg); if (ret < 0) { cerr << "could not create subuser: " << err_msg << std::endl; return -ret; } } break; case OPT_USER_RM: ret = user.remove(user_op, &err_msg); if (ret < 0) { cerr << "could not remove user: " << err_msg << std::endl; return -ret; } output_user_info = false; break; case OPT_USER_ENABLE: case OPT_USER_SUSPEND: case OPT_USER_MODIFY: ret = user.modify(user_op, &err_msg); if (ret < 0) { cerr << "could not modify user: " << err_msg << std::endl; return -ret; } break; case OPT_SUBUSER_CREATE: ret = user.subusers.add(user_op, &err_msg); if (ret < 0) { cerr << "could not create subuser: " << err_msg << std::endl; return -ret; } break; case OPT_SUBUSER_MODIFY: ret = user.subusers.modify(user_op, &err_msg); if (ret < 0) { cerr << "could not modify subuser: " << err_msg << std::endl; return -ret; } break; case OPT_SUBUSER_RM: ret = user.subusers.remove(user_op, &err_msg); if (ret < 0) { cerr << "could not remove subuser: " << err_msg << std::endl; return -ret; } break; case OPT_CAPS_ADD: ret = user.caps.add(user_op, &err_msg); if (ret < 0) { cerr << "could not add caps: " << err_msg << std::endl; return -ret; } break; case OPT_CAPS_RM: ret = user.caps.remove(user_op, &err_msg); if (ret < 0) { cerr << "could not remove caps: " << err_msg << std::endl; return -ret; } break; case OPT_KEY_CREATE: ret = user.keys.add(user_op, &err_msg); if (ret < 0) { cerr << "could not create key: " << err_msg << std::endl; return -ret; } break; case OPT_KEY_RM: ret = user.keys.remove(user_op, &err_msg); if (ret < 0) { cerr << "could not remove key: " << err_msg << std::endl; return -ret; } break; case OPT_PERIOD_PUSH: { RGWEnv env; req_info info(g_ceph_context, &env); info.method = "POST"; info.request_uri = "/admin/realm/period"; map ¶ms = info.args.get_params(); if (!realm_id.empty()) params["realm_id"] = realm_id; if (!realm_name.empty()) params["realm_name"] = realm_name; if (!period_id.empty()) params["period_id"] = period_id; if (!period_epoch.empty()) params["epoch"] = period_epoch; // load the period RGWPeriod period(period_id); int ret = period.init(g_ceph_context, store); if (ret < 0) { cerr << "period init failed: " << cpp_strerror(-ret) << std::endl; return -ret; } // json format into a bufferlist JSONFormatter jf(false); encode_json("period", period, &jf); bufferlist bl; jf.flush(bl); JSONParser p; ret = send_to_remote_or_url(nullptr, url, access_key, secret_key, info, bl, p); if (ret < 0) { cerr << "request failed: " << cpp_strerror(-ret) << std::endl; return -ret; } } return 0; case OPT_PERIOD_UPDATE: { int ret = update_period(realm_id, realm_name, period_id, period_epoch, commit, remote, url, access_key, secret_key, formatter, yes_i_really_mean_it); if (ret < 0) { return -ret; } } return 0; case OPT_PERIOD_COMMIT: { // read realm and staging period RGWRealm realm(realm_id, realm_name); int ret = realm.init(g_ceph_context, store); if (ret < 0) { cerr << "Error initializing realm: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWPeriod period(RGWPeriod::get_staging_id(realm.get_id()), 1); ret = period.init(g_ceph_context, store, realm.get_id()); if (ret < 0) { cerr << "period init failed: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = commit_period(realm, period, remote, url, access_key, secret_key, yes_i_really_mean_it); if (ret < 0) { cerr << "failed to commit period: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("period", period, formatter); formatter->flush(cout); cout << std::endl; } return 0; case OPT_ROLE_CREATE: { if (role_name.empty()) { cerr << "ERROR: role name is empty" << std::endl; return -EINVAL; } if (assume_role_doc.empty()) { cerr << "ERROR: assume role policy document is empty" << std::endl; return -EINVAL; } /* The following two calls will be replaced by read_decode_json or something similar when the code for AWS Policies is in places */ bufferlist bl; int ret = read_input(assume_role_doc, bl); if (ret < 0) { cerr << "ERROR: failed to read input: " << cpp_strerror(-ret) << std::endl; return ret; } JSONParser p; if (!p.parse(bl.c_str(), bl.length())) { cout << "ERROR: failed to parse JSON: " << assume_role_doc << std::endl; return -EINVAL; } string trust_policy = bl.to_str(); RGWRole role(g_ceph_context, store, role_name, path, trust_policy, tenant); ret = role.create(true); if (ret < 0) { return -ret; } show_role_info(role, formatter); return 0; } case OPT_ROLE_DELETE: { if (role_name.empty()) { cerr << "ERROR: empty role name" << std::endl; return -EINVAL; } RGWRole role(g_ceph_context, store, role_name, tenant); ret = role.delete_obj(); if (ret < 0) { return -ret; } cout << "role: " << role_name << " successfully deleted" << std::endl; return 0; } case OPT_ROLE_GET: { if (role_name.empty()) { cerr << "ERROR: empty role name" << std::endl; return -EINVAL; } RGWRole role(g_ceph_context, store, role_name, tenant); ret = role.get(); if (ret < 0) { return -ret; } show_role_info(role, formatter); return 0; } case OPT_ROLE_MODIFY: { if (role_name.empty()) { cerr << "ERROR: role name is empty" << std::endl; return -EINVAL; } if (assume_role_doc.empty()) { cerr << "ERROR: assume role policy document is empty" << std::endl; return -EINVAL; } /* The following two calls will be replaced by read_decode_json or something similar when the code for AWS Policies is in place */ bufferlist bl; int ret = read_input(assume_role_doc, bl); if (ret < 0) { cerr << "ERROR: failed to read input: " << cpp_strerror(-ret) << std::endl; return ret; } JSONParser p; if (!p.parse(bl.c_str(), bl.length())) { cout << "ERROR: failed to parse JSON: " << assume_role_doc << std::endl; return -EINVAL; } string trust_policy = bl.to_str(); RGWRole role(g_ceph_context, store, role_name, tenant); ret = role.get(); if (ret < 0) { return -ret; } role.update_trust_policy(trust_policy); ret = role.update(); if (ret < 0) { return -ret; } cout << "Assume role policy document updated successfully for role: " << role_name << std::endl; return 0; } case OPT_ROLE_LIST: { vector result; ret = RGWRole::get_roles_by_path_prefix(store, g_ceph_context, path_prefix, tenant, result); if (ret < 0) { return -ret; } show_roles_info(result, formatter); return 0; } case OPT_ROLE_POLICY_PUT: { if (role_name.empty()) { cerr << "role name is empty" << std::endl; return -EINVAL; } if (policy_name.empty()) { cerr << "policy name is empty" << std::endl; return -EINVAL; } if (perm_policy_doc.empty()) { cerr << "permission policy document is empty" << std::endl; return -EINVAL; } /* The following two calls will be replaced by read_decode_json or something similar, when code for AWS Policies is in place.*/ bufferlist bl; int ret = read_input(perm_policy_doc, bl); if (ret < 0) { cerr << "ERROR: failed to read input: " << cpp_strerror(-ret) << std::endl; return ret; } JSONParser p; if (!p.parse(bl.c_str(), bl.length())) { cout << "ERROR: failed to parse JSON: " << std::endl; return -EINVAL; } string perm_policy; perm_policy = bl.c_str(); RGWRole role(g_ceph_context, store, role_name, tenant); ret = role.get(); if (ret < 0) { return -ret; } role.set_perm_policy(policy_name, perm_policy); ret = role.update(); if (ret < 0) { return -ret; } cout << "Permission policy attached successfully" << std::endl; return 0; } case OPT_ROLE_POLICY_LIST: { if (role_name.empty()) { cerr << "ERROR: Role name is empty" << std::endl; return -EINVAL; } RGWRole role(g_ceph_context, store, role_name, tenant); ret = role.get(); if (ret < 0) { return -ret; } std::vector policy_names = role.get_role_policy_names(); show_policy_names(policy_names, formatter); return 0; } case OPT_ROLE_POLICY_GET: { if (role_name.empty()) { cerr << "ERROR: role name is empty" << std::endl; return -EINVAL; } if (policy_name.empty()) { cerr << "ERROR: policy name is empty" << std::endl; return -EINVAL; } RGWRole role(g_ceph_context, store, role_name, tenant); int ret = role.get(); if (ret < 0) { return -ret; } string perm_policy; ret = role.get_role_policy(policy_name, perm_policy); if (ret < 0) { return -ret; } show_perm_policy(perm_policy, formatter); return 0; } case OPT_ROLE_POLICY_DELETE: { if (role_name.empty()) { cerr << "ERROR: role name is empty" << std::endl; return -EINVAL; } if (policy_name.empty()) { cerr << "ERROR: policy name is empty" << std::endl; return -EINVAL; } RGWRole role(g_ceph_context, store, role_name, tenant); ret = role.get(); if (ret < 0) { return -ret; } ret = role.delete_policy(policy_name); if (ret < 0) { return -ret; } ret = role.update(); if (ret < 0) { return -ret; } cout << "Policy: " << policy_name << " successfully deleted for role: " << role_name << std::endl; return 0; } default: output_user_info = false; } // output the result of a user operation if (output_user_info) { ret = user.info(info, &err_msg); if (ret < 0) { cerr << "could not fetch user info: " << err_msg << std::endl; return -ret; } show_user_info(info, formatter); } if (opt_cmd == OPT_POLICY) { if (format == "xml") { int ret = RGWBucketAdminOp::dump_s3_policy(store, bucket_op, cout); if (ret < 0) { cerr << "ERROR: failed to get policy: " << cpp_strerror(-ret) << std::endl; return -ret; } } else { int ret = RGWBucketAdminOp::get_policy(store, bucket_op, f); if (ret < 0) { cerr << "ERROR: failed to get policy: " << cpp_strerror(-ret) << std::endl; return -ret; } } } if (opt_cmd == OPT_BUCKET_LIMIT_CHECK) { void *handle; std::list user_ids; metadata_key = "user"; int max = 1000; bool truncated; if (! user_id.empty()) { user_ids.push_back(user_id.id); ret = RGWBucketAdminOp::limit_check(store, bucket_op, user_ids, f, warnings_only); } else { /* list users in groups of max-keys, then perform user-bucket * limit-check on each group */ ret = store->meta_mgr->list_keys_init(metadata_key, &handle); if (ret < 0) { cerr << "ERROR: buckets limit check can't get user metadata_key: " << cpp_strerror(-ret) << std::endl; return -ret; } do { ret = store->meta_mgr->list_keys_next(handle, max, user_ids, &truncated); if (ret < 0 && ret != -ENOENT) { cerr << "ERROR: buckets limit check lists_keys_next(): " << cpp_strerror(-ret) << std::endl; break; } else { /* ok, do the limit checks for this group */ ret = RGWBucketAdminOp::limit_check(store, bucket_op, user_ids, f, warnings_only); if (ret < 0) break; } user_ids.clear(); } while (truncated); store->meta_mgr->list_keys_complete(handle); } return -ret; } /* OPT_BUCKET_LIMIT_CHECK */ if (opt_cmd == OPT_BUCKETS_LIST) { if (bucket_name.empty()) { RGWBucketAdminOp::info(store, bucket_op, f); } else { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_array_section("entries"); bool truncated; int count = 0; if (max_entries < 0) max_entries = 1000; string prefix; string delim; vector result; map common_prefixes; string ns; RGWRados::Bucket target(store, bucket_info); RGWRados::Bucket::List list_op(&target); list_op.params.prefix = prefix; list_op.params.delim = delim; list_op.params.marker = rgw_obj_key(marker); list_op.params.ns = ns; list_op.params.enforce_ns = false; list_op.params.list_versions = true; do { ret = list_op.list_objects(max_entries - count, &result, &common_prefixes, &truncated); if (ret < 0) { cerr << "ERROR: store->list_objects(): " << cpp_strerror(-ret) << std::endl; return -ret; } count += result.size(); for (vector::iterator iter = result.begin(); iter != result.end(); ++iter) { rgw_bucket_dir_entry& entry = *iter; encode_json("entry", entry, formatter); } formatter->flush(cout); } while (truncated && count < max_entries); formatter->close_section(); formatter->flush(cout); } /* have bucket_name */ } /* OPT_BUCKETS_LIST */ if (opt_cmd == OPT_BUCKET_STATS) { bucket_op.set_fetch_stats(true); int r = RGWBucketAdminOp::info(store, bucket_op, f); if (r < 0) { cerr << "failure: " << cpp_strerror(-r) << ": " << err << std::endl; return -r; } } if (opt_cmd == OPT_BUCKET_LINK) { bucket_op.set_bucket_id(bucket_id); string err; int r = RGWBucketAdminOp::link(store, bucket_op, &err); if (r < 0) { cerr << "failure: " << cpp_strerror(-r) << ": " << err << std::endl; return -r; } } if (opt_cmd == OPT_BUCKET_UNLINK) { int r = RGWBucketAdminOp::unlink(store, bucket_op); if (r < 0) { cerr << "failure: " << cpp_strerror(-r) << std::endl; return -r; } } if (opt_cmd == OPT_LOG_LIST) { // filter by date? if (date.size() && date.size() != 10) { cerr << "bad date format for '" << date << "', expect YYYY-MM-DD" << std::endl; return EINVAL; } formatter->reset(); formatter->open_array_section("logs"); RGWAccessHandle h; int r = store->log_list_init(date, &h); if (r == -ENOENT) { // no logs. } else { if (r < 0) { cerr << "log list: error " << r << std::endl; return -r; } while (true) { string name; int r = store->log_list_next(h, &name); if (r == -ENOENT) break; if (r < 0) { cerr << "log list: error " << r << std::endl; return -r; } formatter->dump_string("object", name); } } formatter->close_section(); formatter->flush(cout); cout << std::endl; } if (opt_cmd == OPT_LOG_SHOW || opt_cmd == OPT_LOG_RM) { if (object.empty() && (date.empty() || bucket_name.empty() || bucket_id.empty())) { cerr << "specify an object or a date, bucket and bucket-id" << std::endl; usage(); assert(false); } string oid; if (!object.empty()) { oid = object; } else { oid = date; oid += "-"; oid += bucket_id; oid += "-"; oid += bucket_name; } if (opt_cmd == OPT_LOG_SHOW) { RGWAccessHandle h; int r = store->log_show_init(oid, &h); if (r < 0) { cerr << "error opening log " << oid << ": " << cpp_strerror(-r) << std::endl; return -r; } formatter->reset(); formatter->open_object_section("log"); struct rgw_log_entry entry; // peek at first entry to get bucket metadata r = store->log_show_next(h, &entry); if (r < 0) { cerr << "error reading log " << oid << ": " << cpp_strerror(-r) << std::endl; return -r; } formatter->dump_string("bucket_id", entry.bucket_id); formatter->dump_string("bucket_owner", entry.bucket_owner.to_str()); formatter->dump_string("bucket", entry.bucket); uint64_t agg_time = 0; uint64_t agg_bytes_sent = 0; uint64_t agg_bytes_received = 0; uint64_t total_entries = 0; if (show_log_entries) formatter->open_array_section("log_entries"); do { uint64_t total_time = entry.total_time.sec() * 1000000LL + entry.total_time.usec(); agg_time += total_time; agg_bytes_sent += entry.bytes_sent; agg_bytes_received += entry.bytes_received; total_entries++; if (skip_zero_entries && entry.bytes_sent == 0 && entry.bytes_received == 0) goto next; if (show_log_entries) { rgw_format_ops_log_entry(entry, formatter); formatter->flush(cout); } next: r = store->log_show_next(h, &entry); } while (r > 0); if (r < 0) { cerr << "error reading log " << oid << ": " << cpp_strerror(-r) << std::endl; return -r; } if (show_log_entries) formatter->close_section(); if (show_log_sum) { formatter->open_object_section("log_sum"); formatter->dump_int("bytes_sent", agg_bytes_sent); formatter->dump_int("bytes_received", agg_bytes_received); formatter->dump_int("total_time", agg_time); formatter->dump_int("total_entries", total_entries); formatter->close_section(); } formatter->close_section(); formatter->flush(cout); cout << std::endl; } if (opt_cmd == OPT_LOG_RM) { int r = store->log_remove(oid); if (r < 0) { cerr << "error removing log " << oid << ": " << cpp_strerror(-r) << std::endl; return -r; } } } if (opt_cmd == OPT_POOL_ADD) { if (pool_name.empty()) { cerr << "need to specify pool to add!" << std::endl; usage(); assert(false); } int ret = store->add_bucket_placement(pool); if (ret < 0) cerr << "failed to add bucket placement: " << cpp_strerror(-ret) << std::endl; } if (opt_cmd == OPT_POOL_RM) { if (pool_name.empty()) { cerr << "need to specify pool to remove!" << std::endl; usage(); assert(false); } int ret = store->remove_bucket_placement(pool); if (ret < 0) cerr << "failed to remove bucket placement: " << cpp_strerror(-ret) << std::endl; } if (opt_cmd == OPT_POOLS_LIST) { set pools; int ret = store->list_placement_set(pools); if (ret < 0) { cerr << "could not list placement set: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->reset(); formatter->open_array_section("pools"); for (auto siter = pools.begin(); siter != pools.end(); ++siter) { formatter->open_object_section("pool"); formatter->dump_string("name", siter->to_str()); formatter->close_section(); } formatter->close_section(); formatter->flush(cout); cout << std::endl; } if (opt_cmd == OPT_USAGE_SHOW) { uint64_t start_epoch = 0; uint64_t end_epoch = (uint64_t)-1; int ret; if (!start_date.empty()) { ret = utime_t::parse_date(start_date, &start_epoch, NULL); if (ret < 0) { cerr << "ERROR: failed to parse start date" << std::endl; return 1; } } if (!end_date.empty()) { ret = utime_t::parse_date(end_date, &end_epoch, NULL); if (ret < 0) { cerr << "ERROR: failed to parse end date" << std::endl; return 1; } } ret = RGWUsage::show(store, user_id, start_epoch, end_epoch, show_log_entries, show_log_sum, &categories, f); if (ret < 0) { cerr << "ERROR: failed to show usage" << std::endl; return 1; } } if (opt_cmd == OPT_USAGE_TRIM) { if (user_id.empty() && !yes_i_really_mean_it) { cerr << "usage trim without user specified will remove *all* users data" << std::endl; cerr << "do you really mean it? (requires --yes-i-really-mean-it)" << std::endl; return 1; } int ret; uint64_t start_epoch = 0; uint64_t end_epoch = (uint64_t)-1; if (!start_date.empty()) { ret = utime_t::parse_date(start_date, &start_epoch, NULL); if (ret < 0) { cerr << "ERROR: failed to parse start date" << std::endl; return 1; } } if (!end_date.empty()) { ret = utime_t::parse_date(end_date, &end_epoch, NULL); if (ret < 0) { cerr << "ERROR: failed to parse end date" << std::endl; return 1; } } ret = RGWUsage::trim(store, user_id, start_epoch, end_epoch); if (ret < 0) { cerr << "ERROR: read_usage() returned ret=" << ret << std::endl; return 1; } } if (opt_cmd == OPT_OLH_GET || opt_cmd == OPT_OLH_READLOG) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } if (object.empty()) { cerr << "ERROR: object not specified" << std::endl; return EINVAL; } } if (opt_cmd == OPT_OLH_GET) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWOLHInfo olh; rgw_obj obj(bucket, object); ret = store->get_olh(bucket_info, obj, &olh); if (ret < 0) { cerr << "ERROR: failed reading olh: " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("olh", olh, formatter); formatter->flush(cout); } if (opt_cmd == OPT_OLH_READLOG) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } map > log; bool is_truncated; RGWObjectCtx rctx(store); rgw_obj obj(bucket, object); RGWObjState *state; ret = store->get_obj_state(&rctx, bucket_info, obj, &state, false); /* don't follow olh */ if (ret < 0) { return -ret; } ret = store->bucket_index_read_olh_log(bucket_info, *state, obj, 0, &log, &is_truncated); if (ret < 0) { cerr << "ERROR: failed reading olh: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_object_section("result"); encode_json("is_truncated", is_truncated, formatter); encode_json("log", log, formatter); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_BI_GET) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } rgw_obj obj(bucket, object); if (!object_version.empty()) { obj.key.set_instance(object_version); } rgw_cls_bi_entry entry; ret = store->bi_get(bucket, obj, bi_index_type, &entry); if (ret < 0) { cerr << "ERROR: bi_get(): " << cpp_strerror(-ret) << std::endl; return -ret; } encode_json("entry", entry, formatter); formatter->flush(cout); } if (opt_cmd == OPT_BI_PUT) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } rgw_cls_bi_entry entry; cls_rgw_obj_key key; ret = read_decode_json(infile, entry, &key); if (ret < 0) { return 1; } rgw_obj obj(bucket, key); ret = store->bi_put(bucket, obj, entry); if (ret < 0) { cerr << "ERROR: bi_put(): " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_BI_LIST) { if (bucket_name.empty()) { cerr << "ERROR: bucket name not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } list entries; bool is_truncated; if (max_entries < 0) { max_entries = 1000; } int max_shards = (bucket_info.num_shards > 0 ? bucket_info.num_shards : 1); formatter->open_array_section("entries"); for (int i = 0; i < max_shards; i++) { RGWRados::BucketShard bs(store); int shard_id = (bucket_info.num_shards > 0 ? i : -1); int ret = bs.init(bucket, shard_id); marker.clear(); if (ret < 0) { cerr << "ERROR: bs.init(bucket=" << bucket << ", shard=" << shard_id << "): " << cpp_strerror(-ret) << std::endl; return -ret; } do { entries.clear(); ret = store->bi_list(bs, object, marker, max_entries, &entries, &is_truncated); if (ret < 0) { cerr << "ERROR: bi_list(): " << cpp_strerror(-ret) << std::endl; return -ret; } list::iterator iter; for (iter = entries.begin(); iter != entries.end(); ++iter) { rgw_cls_bi_entry& entry = *iter; encode_json("entry", entry, formatter); marker = entry.idx; } formatter->flush(cout); } while (is_truncated); formatter->flush(cout); } formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_BI_PURGE) { if (bucket_name.empty()) { cerr << "ERROR: bucket name not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWBucketInfo cur_bucket_info; rgw_bucket cur_bucket; ret = init_bucket(tenant, bucket_name, string(), cur_bucket_info, cur_bucket); if (ret < 0) { cerr << "ERROR: could not init current bucket info for bucket_name=" << bucket_name << ": " << cpp_strerror(-ret) << std::endl; return -ret; } if (cur_bucket_info.bucket.bucket_id == bucket_info.bucket.bucket_id && !yes_i_really_mean_it) { cerr << "specified bucket instance points to a current bucket instance" << std::endl; cerr << "do you really mean it? (requires --yes-i-really-mean-it)" << std::endl; return EINVAL; } int max_shards = (bucket_info.num_shards > 0 ? bucket_info.num_shards : 1); for (int i = 0; i < max_shards; i++) { RGWRados::BucketShard bs(store); int shard_id = (bucket_info.num_shards > 0 ? i : -1); int ret = bs.init(bucket, shard_id); if (ret < 0) { cerr << "ERROR: bs.init(bucket=" << bucket << ", shard=" << shard_id << "): " << cpp_strerror(-ret) << std::endl; return -ret; } ret = store->bi_remove(bs); if (ret < 0) { cerr << "ERROR: failed to remove bucket index object: " << cpp_strerror(-ret) << std::endl; return -ret; } } } if (opt_cmd == OPT_OBJECT_RM) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } rgw_obj_key key(object, object_version); ret = rgw_remove_object(store, bucket_info, bucket, key); if (ret < 0) { cerr << "ERROR: object remove returned: " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_OBJECT_REWRITE) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } if (object.empty()) { cerr << "ERROR: object not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } rgw_obj obj(bucket, object); obj.key.set_instance(object_version); bool need_rewrite = true; if (min_rewrite_stripe_size > 0) { ret = check_min_obj_stripe_size(store, bucket_info, obj, min_rewrite_stripe_size, &need_rewrite); if (ret < 0) { ldout(store->ctx(), 0) << "WARNING: check_min_obj_stripe_size failed, r=" << ret << dendl; } } if (need_rewrite) { ret = store->rewrite_obj(bucket_info, obj); if (ret < 0) { cerr << "ERROR: object rewrite returned: " << cpp_strerror(-ret) << std::endl; return -ret; } } else { ldout(store->ctx(), 20) << "skipped object" << dendl; } } if (opt_cmd == OPT_OBJECTS_EXPIRE) { if (!store->process_expire_objects()) { cerr << "ERROR: process_expire_objects() processing returned error." << std::endl; return 1; } } if (opt_cmd == OPT_BUCKET_REWRITE) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } uint64_t start_epoch = 0; uint64_t end_epoch = 0; if (!end_date.empty()) { int ret = utime_t::parse_date(end_date, &end_epoch, NULL); if (ret < 0) { cerr << "ERROR: failed to parse end date" << std::endl; return EINVAL; } } if (!start_date.empty()) { int ret = utime_t::parse_date(start_date, &start_epoch, NULL); if (ret < 0) { cerr << "ERROR: failed to parse start date" << std::endl; return EINVAL; } } bool is_truncated = true; rgw_obj_index_key marker; string prefix; formatter->open_object_section("result"); formatter->dump_string("bucket", bucket_name); formatter->open_array_section("objects"); while (is_truncated) { map result; int r = store->cls_bucket_list_ordered(bucket_info, RGW_NO_SHARD, marker, prefix, 1000, true, result, &is_truncated, &marker, bucket_object_check_filter); if (r < 0 && r != -ENOENT) { cerr << "ERROR: failed operation r=" << r << std::endl; } if (r == -ENOENT) break; map::iterator iter; for (iter = result.begin(); iter != result.end(); ++iter) { rgw_obj_key key = iter->second.key; rgw_bucket_dir_entry& entry = iter->second; formatter->open_object_section("object"); formatter->dump_string("name", key.name); formatter->dump_string("instance", key.instance); formatter->dump_int("size", entry.meta.size); utime_t ut(entry.meta.mtime); ut.gmtime(formatter->dump_stream("mtime")); if ((entry.meta.size < min_rewrite_size) || (entry.meta.size > max_rewrite_size) || (start_epoch > 0 && start_epoch > (uint64_t)ut.sec()) || (end_epoch > 0 && end_epoch < (uint64_t)ut.sec())) { formatter->dump_string("status", "Skipped"); } else { rgw_obj obj(bucket, key); bool need_rewrite = true; if (min_rewrite_stripe_size > 0) { r = check_min_obj_stripe_size(store, bucket_info, obj, min_rewrite_stripe_size, &need_rewrite); if (r < 0) { ldout(store->ctx(), 0) << "WARNING: check_min_obj_stripe_size failed, r=" << r << dendl; } } if (!need_rewrite) { formatter->dump_string("status", "Skipped"); } else { r = store->rewrite_obj(bucket_info, obj); if (r == 0) { formatter->dump_string("status", "Success"); } else { formatter->dump_string("status", cpp_strerror(-r)); } } } formatter->dump_int("flags", entry.flags); formatter->close_section(); formatter->flush(cout); } } formatter->close_section(); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_BUCKET_RESHARD) { rgw_bucket bucket; RGWBucketInfo bucket_info; map attrs; int ret = check_reshard_bucket_params(store, bucket_name, tenant, bucket_id, num_shards_specified, num_shards, yes_i_really_mean_it, bucket, bucket_info, attrs); if (ret < 0) { return ret; } RGWBucketReshard br(store, bucket_info, attrs); #define DEFAULT_RESHARD_MAX_ENTRIES 1000 if (max_entries < 1) { max_entries = DEFAULT_RESHARD_MAX_ENTRIES; } return br.execute(num_shards, max_entries, verbose, &cout, formatter); } if (opt_cmd == OPT_RESHARD_ADD) { rgw_bucket bucket; RGWBucketInfo bucket_info; map attrs; int ret = check_reshard_bucket_params(store, bucket_name, tenant, bucket_id, num_shards_specified, num_shards, yes_i_really_mean_it, bucket, bucket_info, attrs); if (ret < 0) { return ret; } int num_source_shards = (bucket_info.num_shards > 0 ? bucket_info.num_shards : 1); RGWReshard reshard(store); cls_rgw_reshard_entry entry; entry.time = real_clock::now(); entry.tenant = tenant; entry.bucket_name = bucket_name; entry.bucket_id = bucket_info.bucket.bucket_id; entry.old_num_shards = num_source_shards; entry.new_num_shards = num_shards; return reshard.add(entry); } if (opt_cmd == OPT_RESHARD_LIST) { list entries; int ret; int count = 0; if (max_entries < 0) { max_entries = 1000; } int num_logshards = store->ctx()->_conf->rgw_reshard_num_logs; RGWReshard reshard(store); formatter->open_array_section("reshard"); for (int i = 0; i < num_logshards; i++) { bool is_truncated = true; string marker; do { entries.clear(); ret = reshard.list(i, marker, max_entries, entries, &is_truncated); if (ret < 0) { cerr << "Error listing resharding buckets: " << cpp_strerror(-ret) << std::endl; return ret; } for (auto iter=entries.begin(); iter != entries.end(); ++iter) { cls_rgw_reshard_entry& entry = *iter; encode_json("entry", entry, formatter); entry.get_key(&marker); } count += entries.size(); formatter->flush(cout); } while (is_truncated && count < max_entries); if (count >= max_entries) { break; } } formatter->close_section(); formatter->flush(cout); return 0; } if (opt_cmd == OPT_RESHARD_STATUS) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } rgw_bucket bucket; RGWBucketInfo bucket_info; map attrs; ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket, &attrs); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWBucketReshard br(store, bucket_info, attrs); list status; int r = br.get_status(&status); if (r < 0) { cerr << "ERROR: could not get resharding status for bucket " << bucket_name << std::endl; return -r; } encode_json("status", status, formatter); formatter->flush(cout); } if (opt_cmd == OPT_RESHARD_PROCESS) { RGWReshard reshard(store, true, &cout); int ret = reshard.process_all_logshards(); if (ret < 0) { cerr << "ERROR: failed to process reshard logs, error=" << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_RESHARD_CANCEL) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } rgw_bucket bucket; RGWBucketInfo bucket_info; map attrs; ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket, &attrs); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWBucketReshard br(store, bucket_info, attrs); int ret = br.cancel(); if (ret < 0) { if (ret == -EBUSY) { cerr << "There is ongoing resharding, please retry after " << g_conf->rgw_reshard_bucket_lock_duration << " seconds " << std::endl; } else { cerr << "Error canceling bucket " << bucket_name << " resharding: " << cpp_strerror(-ret) << std::endl; } return ret; } RGWReshard reshard(store); cls_rgw_reshard_entry entry; //entry.tenant = tenant; entry.bucket_name = bucket_name; //entry.bucket_id = bucket_id; ret = reshard.remove(entry); if (ret < 0 && ret != -ENOENT) { cerr << "Error in getting bucket " << bucket_name << ": " << cpp_strerror(-ret) << std::endl; return ret; } } if (opt_cmd == OPT_OBJECT_UNLINK) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } list oid_list; rgw_obj_key key(object, object_version); rgw_obj_index_key index_key; key.get_index_key(&index_key); oid_list.push_back(index_key); ret = store->remove_objs_from_index(bucket_info, oid_list); if (ret < 0) { cerr << "ERROR: remove_obj_from_index() returned error: " << cpp_strerror(-ret) << std::endl; return 1; } } if (opt_cmd == OPT_OBJECT_STAT) { RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } rgw_obj obj(bucket, object); obj.key.set_instance(object_version); uint64_t obj_size; map attrs; RGWObjectCtx obj_ctx(store); RGWRados::Object op_target(store, bucket_info, obj_ctx, obj); RGWRados::Object::Read read_op(&op_target); read_op.params.attrs = &attrs; read_op.params.obj_size = &obj_size; ret = read_op.prepare(); if (ret < 0) { cerr << "ERROR: failed to stat object, returned error: " << cpp_strerror(-ret) << std::endl; return 1; } formatter->open_object_section("object_metadata"); formatter->dump_string("name", object); formatter->dump_unsigned("size", obj_size); map::iterator iter; map other_attrs; for (iter = attrs.begin(); iter != attrs.end(); ++iter) { bufferlist& bl = iter->second; bool handled = false; if (iter->first == RGW_ATTR_MANIFEST) { handled = decode_dump("manifest", bl, formatter); } else if (iter->first == RGW_ATTR_ACL) { handled = decode_dump("policy", bl, formatter); } else if (iter->first == RGW_ATTR_ID_TAG) { handled = dump_string("tag", bl, formatter); } else if (iter->first == RGW_ATTR_ETAG) { handled = dump_string("etag", bl, formatter); } if (!handled) other_attrs[iter->first] = bl; } formatter->open_object_section("attrs"); for (iter = other_attrs.begin(); iter != other_attrs.end(); ++iter) { dump_string(iter->first.c_str(), iter->second, formatter); } formatter->close_section(); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_BUCKET_CHECK) { if (check_head_obj_locator) { if (bucket_name.empty()) { cerr << "ERROR: need to specify bucket name" << std::endl; return EINVAL; } do_check_object_locator(tenant, bucket_name, fix, remove_bad, formatter); } else { RGWBucketAdminOp::check_index(store, bucket_op, f); } } if (opt_cmd == OPT_BUCKET_RM) { if (inconsistent_index == false) { RGWBucketAdminOp::remove_bucket(store, bucket_op, bypass_gc, true); } else { if (!yes_i_really_mean_it) { cerr << "using --inconsistent_index can corrupt the bucket index " << std::endl << "do you really mean it? (requires --yes-i-really-mean-it)" << std::endl; return 1; } RGWBucketAdminOp::remove_bucket(store, bucket_op, bypass_gc, false); } } if (opt_cmd == OPT_GC_LIST) { int index = 0; bool truncated; formatter->open_array_section("entries"); do { list result; int ret = store->list_gc_objs(&index, marker, 1000, !include_all, result, &truncated); if (ret < 0) { cerr << "ERROR: failed to list objs: " << cpp_strerror(-ret) << std::endl; return 1; } list::iterator iter; for (iter = result.begin(); iter != result.end(); ++iter) { cls_rgw_gc_obj_info& info = *iter; formatter->open_object_section("chain_info"); formatter->dump_string("tag", info.tag); formatter->dump_stream("time") << info.time; formatter->open_array_section("objs"); list::iterator liter; cls_rgw_obj_chain& chain = info.chain; for (liter = chain.objs.begin(); liter != chain.objs.end(); ++liter) { cls_rgw_obj& obj = *liter; encode_json("obj", obj, formatter); } formatter->close_section(); // objs formatter->close_section(); // obj_chain formatter->flush(cout); } } while (truncated); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_GC_PROCESS) { int ret = store->process_gc(); if (ret < 0) { cerr << "ERROR: gc processing returned error: " << cpp_strerror(-ret) << std::endl; return 1; } } if (opt_cmd == OPT_LC_LIST) { formatter->open_array_section("lifecycle_list"); map bucket_lc_map; string marker; #define MAX_LC_LIST_ENTRIES 100 if (max_entries < 0) { max_entries = MAX_LC_LIST_ENTRIES; } do { int ret = store->list_lc_progress(marker, max_entries, &bucket_lc_map); if (ret < 0) { cerr << "ERROR: failed to list objs: " << cpp_strerror(-ret) << std::endl; return 1; } map::iterator iter; for (iter = bucket_lc_map.begin(); iter != bucket_lc_map.end(); ++iter) { formatter->open_object_section("bucket_lc_info"); formatter->dump_string("bucket", iter->first); string lc_status = LC_STATUS[iter->second]; formatter->dump_string("status", lc_status); formatter->close_section(); // objs formatter->flush(cout); marker = iter->first; } } while (!bucket_lc_map.empty()); formatter->close_section(); //lifecycle list formatter->flush(cout); } if (opt_cmd == OPT_LC_PROCESS) { int ret = store->process_lc(); if (ret < 0) { cerr << "ERROR: lc processing returned error: " << cpp_strerror(-ret) << std::endl; return 1; } } if (opt_cmd == OPT_ORPHANS_FIND) { if (!yes_i_really_mean_it) { cerr << "accidental removal of active objects can not be reversed; " << "do you really mean it? (requires --yes-i-really-mean-it)" << std::endl; return EINVAL; } RGWOrphanSearch search(store, max_concurrent_ios, orphan_stale_secs); if (job_id.empty()) { cerr << "ERROR: --job-id not specified" << std::endl; return EINVAL; } if (pool_name.empty()) { cerr << "ERROR: --pool not specified" << std::endl; return EINVAL; } RGWOrphanSearchInfo info; info.pool = pool; info.job_name = job_id; info.num_shards = num_shards; int ret = search.init(job_id, &info); if (ret < 0) { cerr << "could not init search, ret=" << ret << std::endl; return -ret; } ret = search.run(); if (ret < 0) { return -ret; } } if (opt_cmd == OPT_ORPHANS_FINISH) { RGWOrphanSearch search(store, max_concurrent_ios, orphan_stale_secs); if (job_id.empty()) { cerr << "ERROR: --job-id not specified" << std::endl; return EINVAL; } int ret = search.init(job_id, NULL); if (ret < 0) { if (ret == -ENOENT) { cerr << "job not found" << std::endl; } return -ret; } ret = search.finish(); if (ret < 0) { return -ret; } } if (opt_cmd == OPT_ORPHANS_LIST_JOBS){ RGWOrphanStore orphan_store(store); int ret = orphan_store.init(); if (ret < 0){ cerr << "connection to cluster failed!" << std::endl; return -ret; } map m; ret = orphan_store.list_jobs(m); if (ret < 0) { cerr << "job list failed" << std::endl; return -ret; } formatter->open_array_section("entries"); for (const auto &it: m){ if (!extra_info){ formatter->dump_string("job-id",it.first); } else { encode_json("orphan_search_state", it.second, formatter); } } formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_USER_CHECK) { check_bad_user_bucket_mapping(store, user_id, fix); } if (opt_cmd == OPT_USER_STATS) { if (user_id.empty()) { cerr << "ERROR: uid not specified" << std::endl; return EINVAL; } string user_str = user_id.to_str(); if (reset_stats) { if (!bucket_name.empty()){ cerr << "ERROR: recalculate doesn't work on buckets" << std::endl; return EINVAL; } ret = store->cls_user_reset_stats(user_str); if (ret < 0) { cerr << "ERROR: could not clear user stats: " << cpp_strerror(-ret) << std::endl; return -ret; } } if (sync_stats) { if (!bucket_name.empty()) { int ret = rgw_bucket_sync_user_stats(store, tenant, bucket_name); if (ret < 0) { cerr << "ERROR: could not sync bucket stats: " << cpp_strerror(-ret) << std::endl; return -ret; } } else { int ret = rgw_user_sync_all_stats(store, user_id); if (ret < 0) { cerr << "ERROR: failed to sync user stats: " << cpp_strerror(-ret) << std::endl; return -ret; } } } cls_user_header header; int ret = store->cls_user_get_header(user_str, &header); if (ret < 0) { if (ret == -ENOENT) { /* in case of ENOENT */ cerr << "User has not been initialized or user does not exist" << std::endl; } else { cerr << "ERROR: can't read user: " << cpp_strerror(ret) << std::endl; } return -ret; } encode_json("header", header, formatter); formatter->flush(cout); } if (opt_cmd == OPT_METADATA_GET) { int ret = store->meta_mgr->get(metadata_key, formatter); if (ret < 0) { cerr << "ERROR: can't get key: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->flush(cout); } if (opt_cmd == OPT_METADATA_PUT) { bufferlist bl; int ret = read_input(infile, bl); if (ret < 0) { cerr << "ERROR: failed to read input: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = store->meta_mgr->put(metadata_key, bl, RGWMetadataHandler::APPLY_ALWAYS); if (ret < 0) { cerr << "ERROR: can't put key: " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_METADATA_RM) { int ret = store->meta_mgr->remove(metadata_key); if (ret < 0) { cerr << "ERROR: can't remove key: " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_METADATA_LIST || opt_cmd == OPT_USER_LIST) { if (opt_cmd == OPT_USER_LIST) { metadata_key = "user"; } void *handle; int max = 1000; int ret = store->meta_mgr->list_keys_init(metadata_key, marker, &handle); if (ret < 0) { cerr << "ERROR: can't get key: " << cpp_strerror(-ret) << std::endl; return -ret; } bool truncated; uint64_t count = 0; if (max_entries_specified) { formatter->open_object_section("result"); } formatter->open_array_section("keys"); uint64_t left; do { list keys; left = (max_entries_specified ? max_entries - count : max); ret = store->meta_mgr->list_keys_next(handle, left, keys, &truncated); if (ret < 0 && ret != -ENOENT) { cerr << "ERROR: lists_keys_next(): " << cpp_strerror(-ret) << std::endl; return -ret; } if (ret != -ENOENT) { for (list::iterator iter = keys.begin(); iter != keys.end(); ++iter) { formatter->dump_string("key", *iter); ++count; } formatter->flush(cout); } } while (truncated && left > 0); formatter->close_section(); if (max_entries_specified) { encode_json("truncated", truncated, formatter); encode_json("count", count, formatter); if (truncated) { encode_json("marker", store->meta_mgr->get_marker(handle), formatter); } formatter->close_section(); } formatter->flush(cout); store->meta_mgr->list_keys_complete(handle); } if (opt_cmd == OPT_MDLOG_LIST) { utime_t start_time, end_time; int ret = parse_date_str(start_date, start_time); if (ret < 0) return -ret; ret = parse_date_str(end_date, end_time); if (ret < 0) return -ret; int i = (specified_shard_id ? shard_id : 0); if (period_id.empty()) { int ret = read_current_period_id(store, realm_id, realm_name, &period_id); if (ret < 0) { return -ret; } std::cerr << "No --period given, using current period=" << period_id << std::endl; } RGWMetadataLog *meta_log = store->meta_mgr->get_log(period_id); formatter->open_array_section("entries"); for (; i < g_ceph_context->_conf->rgw_md_log_max_shards; i++) { void *handle; list entries; meta_log->init_list_entries(i, start_time.to_real_time(), end_time.to_real_time(), marker, &handle); bool truncated; do { int ret = meta_log->list_entries(handle, 1000, entries, NULL, &truncated); if (ret < 0) { cerr << "ERROR: meta_log->list_entries(): " << cpp_strerror(-ret) << std::endl; return -ret; } for (list::iterator iter = entries.begin(); iter != entries.end(); ++iter) { cls_log_entry& entry = *iter; store->meta_mgr->dump_log_entry(entry, formatter); } formatter->flush(cout); } while (truncated); meta_log->complete_list_entries(handle); if (specified_shard_id) break; } formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_MDLOG_STATUS) { int i = (specified_shard_id ? shard_id : 0); if (period_id.empty()) { int ret = read_current_period_id(store, realm_id, realm_name, &period_id); if (ret < 0) { return -ret; } std::cerr << "No --period given, using current period=" << period_id << std::endl; } RGWMetadataLog *meta_log = store->meta_mgr->get_log(period_id); formatter->open_array_section("entries"); for (; i < g_ceph_context->_conf->rgw_md_log_max_shards; i++) { RGWMetadataLogInfo info; meta_log->get_info(i, &info); ::encode_json("info", info, formatter); if (specified_shard_id) break; } formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_MDLOG_AUTOTRIM) { // need a full history for purging old mdlog periods store->meta_mgr->init_oldest_log_period(); RGWCoroutinesManager crs(store->ctx(), store->get_cr_registry()); RGWHTTPManager http(store->ctx(), crs.get_completion_mgr()); int ret = http.set_threaded(); if (ret < 0) { cerr << "failed to initialize http client with " << cpp_strerror(ret) << std::endl; return -ret; } auto num_shards = g_conf->rgw_md_log_max_shards; ret = crs.run(create_admin_meta_log_trim_cr(store, &http, num_shards)); if (ret < 0) { cerr << "automated mdlog trim failed with " << cpp_strerror(ret) << std::endl; return -ret; } } if (opt_cmd == OPT_MDLOG_TRIM) { utime_t start_time, end_time; if (!specified_shard_id) { cerr << "ERROR: shard-id must be specified for trim operation" << std::endl; return EINVAL; } int ret = parse_date_str(start_date, start_time); if (ret < 0) return -ret; ret = parse_date_str(end_date, end_time); if (ret < 0) return -ret; if (period_id.empty()) { std::cerr << "missing --period argument" << std::endl; return EINVAL; } RGWMetadataLog *meta_log = store->meta_mgr->get_log(period_id); ret = meta_log->trim(shard_id, start_time.to_real_time(), end_time.to_real_time(), start_marker, end_marker); if (ret < 0) { cerr << "ERROR: meta_log->trim(): " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_SYNC_STATUS) { sync_status(formatter); } if (opt_cmd == OPT_METADATA_SYNC_STATUS) { RGWMetaSyncStatusManager sync(store, store->get_async_rados()); int ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } rgw_meta_sync_status sync_status; ret = sync.read_sync_status(&sync_status); if (ret < 0) { cerr << "ERROR: sync.read_sync_status() returned ret=" << ret << std::endl; return -ret; } formatter->open_object_section("summary"); encode_json("sync_status", sync_status, formatter); uint64_t full_total = 0; uint64_t full_complete = 0; for (auto marker_iter : sync_status.sync_markers) { full_total += marker_iter.second.total_entries; if (marker_iter.second.state == rgw_meta_sync_marker::SyncState::FullSync) { full_complete += marker_iter.second.pos; } else { full_complete += marker_iter.second.total_entries; } } formatter->open_object_section("full_sync"); encode_json("total", full_total, formatter); encode_json("complete", full_complete, formatter); formatter->close_section(); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_METADATA_SYNC_INIT) { RGWMetaSyncStatusManager sync(store, store->get_async_rados()); int ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } ret = sync.init_sync_status(); if (ret < 0) { cerr << "ERROR: sync.init_sync_status() returned ret=" << ret << std::endl; return -ret; } } if (opt_cmd == OPT_METADATA_SYNC_RUN) { RGWMetaSyncStatusManager sync(store, store->get_async_rados()); int ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } ret = sync.run(); if (ret < 0) { cerr << "ERROR: sync.run() returned ret=" << ret << std::endl; return -ret; } } if (opt_cmd == OPT_DATA_SYNC_STATUS) { if (source_zone.empty()) { cerr << "ERROR: source zone not specified" << std::endl; return EINVAL; } RGWDataSyncStatusManager sync(store, store->get_async_rados(), source_zone); int ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } rgw_data_sync_status sync_status; if (specified_shard_id) { set pending_buckets; set recovering_buckets; rgw_data_sync_marker sync_marker; ret = sync.read_shard_status(shard_id, pending_buckets, recovering_buckets, &sync_marker, max_entries_specified ? max_entries : 20); if (ret < 0 && ret != -ENOENT) { cerr << "ERROR: sync.read_shard_status() returned ret=" << ret << std::endl; return -ret; } formatter->open_object_section("summary"); encode_json("shard_id", shard_id, formatter); encode_json("marker", sync_marker, formatter); encode_json("pending_buckets", pending_buckets, formatter); encode_json("recovering_buckets", recovering_buckets, formatter); formatter->close_section(); formatter->flush(cout); } else { ret = sync.read_sync_status(&sync_status); if (ret < 0 && ret != -ENOENT) { cerr << "ERROR: sync.read_sync_status() returned ret=" << ret << std::endl; return -ret; } formatter->open_object_section("summary"); encode_json("sync_status", sync_status, formatter); uint64_t full_total = 0; uint64_t full_complete = 0; for (auto marker_iter : sync_status.sync_markers) { full_total += marker_iter.second.total_entries; if (marker_iter.second.state == rgw_meta_sync_marker::SyncState::FullSync) { full_complete += marker_iter.second.pos; } else { full_complete += marker_iter.second.total_entries; } } formatter->open_object_section("full_sync"); encode_json("total", full_total, formatter); encode_json("complete", full_complete, formatter); formatter->close_section(); formatter->close_section(); formatter->flush(cout); } } if (opt_cmd == OPT_DATA_SYNC_INIT) { if (source_zone.empty()) { cerr << "ERROR: source zone not specified" << std::endl; return EINVAL; } RGWDataSyncStatusManager sync(store, store->get_async_rados(), source_zone); int ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } ret = sync.init_sync_status(); if (ret < 0) { cerr << "ERROR: sync.init_sync_status() returned ret=" << ret << std::endl; return -ret; } } if (opt_cmd == OPT_DATA_SYNC_RUN) { if (source_zone.empty()) { cerr << "ERROR: source zone not specified" << std::endl; return EINVAL; } RGWSyncModuleInstanceRef sync_module; int ret = store->get_sync_modules_manager()->create_instance(g_ceph_context, store->get_zone().tier_type, store->get_zone_params().tier_config, &sync_module); if (ret < 0) { lderr(cct) << "ERROR: failed to init sync module instance, ret=" << ret << dendl; return ret; } RGWDataSyncStatusManager sync(store, store->get_async_rados(), source_zone, sync_module); ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } ret = sync.run(); if (ret < 0) { cerr << "ERROR: sync.run() returned ret=" << ret << std::endl; return -ret; } } if (opt_cmd == OPT_BUCKET_SYNC_INIT) { if (source_zone.empty()) { cerr << "ERROR: source zone not specified" << std::endl; return EINVAL; } if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } rgw_bucket bucket; int ret = init_bucket_for_sync(tenant, bucket_name, bucket_id, bucket); if (ret < 0) { return -ret; } RGWBucketSyncStatusManager sync(store, source_zone, bucket); ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } ret = sync.init_sync_status(); if (ret < 0) { cerr << "ERROR: sync.init_sync_status() returned ret=" << ret << std::endl; return -ret; } } if ((opt_cmd == OPT_BUCKET_SYNC_DISABLE) || (opt_cmd == OPT_BUCKET_SYNC_ENABLE)) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } if (ret < 0) { cerr << "could not init realm " << ": " << cpp_strerror(-ret) << std::endl; return ret; } RGWPeriod period; ret = period.init(g_ceph_context, store, realm_id, realm_name, true); if (ret < 0) { cerr << "failed to init period " << ": " << cpp_strerror(-ret) << std::endl; return ret; } if (!store->is_meta_master()) { cerr << "failed to update bucket sync: only allowed on meta master zone " << std::endl; cerr << period.get_master_zone() << " | " << period.get_realm() << std::endl; return EINVAL; } rgw_obj obj(bucket, object); ret = set_bucket_sync_enabled(store, opt_cmd, tenant, bucket_name); if (ret < 0) return -ret; } if (opt_cmd == OPT_BUCKET_SYNC_STATUS) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; rgw_bucket bucket; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { return -ret; } bucket_sync_status(store, bucket_info, source_zone, std::cout); } if (opt_cmd == OPT_BUCKET_SYNC_MARKERS) { if (source_zone.empty()) { cerr << "ERROR: source zone not specified" << std::endl; return EINVAL; } if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } rgw_bucket bucket; int ret = init_bucket_for_sync(tenant, bucket_name, bucket_id, bucket); if (ret < 0) { return -ret; } RGWBucketSyncStatusManager sync(store, source_zone, bucket); ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } ret = sync.read_sync_status(); if (ret < 0) { cerr << "ERROR: sync.read_sync_status() returned ret=" << ret << std::endl; return -ret; } map& sync_status = sync.get_sync_status(); encode_json("sync_status", sync_status, formatter); formatter->flush(cout); } if (opt_cmd == OPT_BUCKET_SYNC_RUN) { if (source_zone.empty()) { cerr << "ERROR: source zone not specified" << std::endl; return EINVAL; } if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } rgw_bucket bucket; int ret = init_bucket_for_sync(tenant, bucket_name, bucket_id, bucket); if (ret < 0) { return -ret; } RGWBucketSyncStatusManager sync(store, source_zone, bucket); ret = sync.init(); if (ret < 0) { cerr << "ERROR: sync.init() returned ret=" << ret << std::endl; return -ret; } ret = sync.run(); if (ret < 0) { cerr << "ERROR: sync.run() returned ret=" << ret << std::endl; return -ret; } } if (opt_cmd == OPT_BILOG_LIST) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_array_section("entries"); bool truncated; int count = 0; if (max_entries < 0) max_entries = 1000; do { list entries; ret = store->list_bi_log_entries(bucket_info, shard_id, marker, max_entries - count, entries, &truncated); if (ret < 0) { cerr << "ERROR: list_bi_log_entries(): " << cpp_strerror(-ret) << std::endl; return -ret; } count += entries.size(); for (list::iterator iter = entries.begin(); iter != entries.end(); ++iter) { rgw_bi_log_entry& entry = *iter; encode_json("entry", entry, formatter); marker = entry.id; } formatter->flush(cout); } while (truncated && count < max_entries); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_SYNC_ERROR_LIST) { if (max_entries < 0) { max_entries = 1000; } bool truncated; utime_t start_time, end_time; int ret = parse_date_str(start_date, start_time); if (ret < 0) return -ret; ret = parse_date_str(end_date, end_time); if (ret < 0) return -ret; if (shard_id < 0) { shard_id = 0; } formatter->open_array_section("entries"); for (; shard_id < ERROR_LOGGER_SHARDS; ++shard_id) { formatter->open_object_section("shard"); encode_json("shard_id", shard_id, formatter); formatter->open_array_section("entries"); int count = 0; string oid = RGWSyncErrorLogger::get_shard_oid(RGW_SYNC_ERROR_LOG_SHARD_PREFIX, shard_id); do { list entries; ret = store->time_log_list(oid, start_time.to_real_time(), end_time.to_real_time(), max_entries - count, entries, marker, &marker, &truncated); if (ret == -ENOENT) { break; } if (ret < 0) { cerr << "ERROR: store->time_log_list(): " << cpp_strerror(-ret) << std::endl; return -ret; } count += entries.size(); for (auto& cls_entry : entries) { rgw_sync_error_info log_entry; auto iter = cls_entry.data.begin(); try { ::decode(log_entry, iter); } catch (buffer::error& err) { cerr << "ERROR: failed to decode log entry" << std::endl; continue; } formatter->open_object_section("entry"); encode_json("id", cls_entry.id, formatter); encode_json("section", cls_entry.section, formatter); encode_json("name", cls_entry.name, formatter); encode_json("timestamp", cls_entry.timestamp, formatter); encode_json("info", log_entry, formatter); formatter->close_section(); formatter->flush(cout); } } while (truncated && count < max_entries); formatter->close_section(); formatter->close_section(); if (specified_shard_id) { break; } } formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_SYNC_ERROR_TRIM) { utime_t start_time, end_time; int ret = parse_date_str(start_date, start_time); if (ret < 0) return -ret; ret = parse_date_str(end_date, end_time); if (ret < 0) return -ret; if (shard_id < 0) { shard_id = 0; } for (; shard_id < ERROR_LOGGER_SHARDS; ++shard_id) { ret = trim_sync_error_log(shard_id, start_time.to_real_time(), end_time.to_real_time(), start_marker, end_marker, trim_delay_ms); if (ret < 0) { cerr << "ERROR: sync error trim: " << cpp_strerror(-ret) << std::endl; return -ret; } if (specified_shard_id) { break; } } } if (opt_cmd == OPT_BILOG_TRIM) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } ret = store->trim_bi_log_entries(bucket_info, shard_id, start_marker, end_marker); if (ret < 0) { cerr << "ERROR: trim_bi_log_entries(): " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_BILOG_STATUS) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } map markers; ret = store->get_bi_log_status(bucket_info, shard_id, markers); if (ret < 0) { cerr << "ERROR: get_bi_log_status(): " << cpp_strerror(-ret) << std::endl; return -ret; } formatter->open_object_section("entries"); encode_json("markers", markers, formatter); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_BILOG_AUTOTRIM) { RGWCoroutinesManager crs(store->ctx(), store->get_cr_registry()); RGWHTTPManager http(store->ctx(), crs.get_completion_mgr()); int ret = http.set_threaded(); if (ret < 0) { cerr << "failed to initialize http client with " << cpp_strerror(ret) << std::endl; return -ret; } rgw::BucketTrimConfig config; configure_bucket_trim(store->ctx(), config); rgw::BucketTrimManager trim(store, config); ret = trim.init(); if (ret < 0) { cerr << "trim manager init failed with " << cpp_strerror(ret) << std::endl; return -ret; } ret = crs.run(trim.create_admin_bucket_trim_cr(&http)); if (ret < 0) { cerr << "automated bilog trim failed with " << cpp_strerror(ret) << std::endl; return -ret; } } if (opt_cmd == OPT_DATALOG_LIST) { formatter->open_array_section("entries"); bool truncated; int count = 0; if (max_entries < 0) max_entries = 1000; utime_t start_time, end_time; int ret = parse_date_str(start_date, start_time); if (ret < 0) return -ret; ret = parse_date_str(end_date, end_time); if (ret < 0) return -ret; RGWDataChangesLog *log = store->data_log; RGWDataChangesLog::LogMarker marker; do { list entries; ret = log->list_entries(start_time.to_real_time(), end_time.to_real_time(), max_entries - count, entries, marker, &truncated); if (ret < 0) { cerr << "ERROR: list_bi_log_entries(): " << cpp_strerror(-ret) << std::endl; return -ret; } count += entries.size(); for (list::iterator iter = entries.begin(); iter != entries.end(); ++iter) { rgw_data_change_log_entry& entry = *iter; if (!extra_info) { encode_json("entry", entry.entry, formatter); } else { encode_json("entry", entry, formatter); } } formatter->flush(cout); } while (truncated && count < max_entries); formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_DATALOG_STATUS) { RGWDataChangesLog *log = store->data_log; int i = (specified_shard_id ? shard_id : 0); formatter->open_array_section("entries"); for (; i < g_ceph_context->_conf->rgw_data_log_num_shards; i++) { list entries; RGWDataChangesLogInfo info; log->get_info(i, &info); ::encode_json("info", info, formatter); if (specified_shard_id) break; } formatter->close_section(); formatter->flush(cout); } if (opt_cmd == OPT_DATALOG_TRIM) { utime_t start_time, end_time; int ret = parse_date_str(start_date, start_time); if (ret < 0) return -ret; ret = parse_date_str(end_date, end_time); if (ret < 0) return -ret; RGWDataChangesLog *log = store->data_log; ret = log->trim_entries(start_time.to_real_time(), end_time.to_real_time(), start_marker, end_marker); if (ret < 0) { cerr << "ERROR: trim_entries(): " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_OPSTATE_LIST) { RGWOpState oc(store); int max = 1000; void *handle; oc.init_list_entries(client_id, op_id, object, &handle); list entries; bool done; formatter->open_array_section("entries"); do { int ret = oc.list_entries(handle, max, entries, &done); if (ret < 0) { cerr << "oc.list_entries returned " << cpp_strerror(-ret) << std::endl; oc.finish_list_entries(handle); return -ret; } for (list::iterator iter = entries.begin(); iter != entries.end(); ++iter) { oc.dump_entry(*iter, formatter); } formatter->flush(cout); } while (!done); formatter->close_section(); formatter->flush(cout); oc.finish_list_entries(handle); } if (opt_cmd == OPT_OPSTATE_SET || opt_cmd == OPT_OPSTATE_RENEW) { RGWOpState oc(store); RGWOpState::OpState state; if (object.empty() || client_id.empty() || op_id.empty()) { cerr << "ERROR: need to specify client_id, op_id, and object" << std::endl; return EINVAL; } if (state_str.empty()) { cerr << "ERROR: state was not specified" << std::endl; return EINVAL; } int ret = oc.state_from_str(state_str, &state); if (ret < 0) { cerr << "ERROR: invalid state: " << state_str << std::endl; return -ret; } if (opt_cmd == OPT_OPSTATE_SET) { ret = oc.set_state(client_id, op_id, object, state); if (ret < 0) { cerr << "ERROR: failed to set state: " << cpp_strerror(-ret) << std::endl; return -ret; } } else { ret = oc.renew_state(client_id, op_id, object, state); if (ret < 0) { cerr << "ERROR: failed to renew state: " << cpp_strerror(-ret) << std::endl; return -ret; } } } if (opt_cmd == OPT_OPSTATE_RM) { RGWOpState oc(store); if (object.empty() || client_id.empty() || op_id.empty()) { cerr << "ERROR: need to specify client_id, op_id, and object" << std::endl; return EINVAL; } ret = oc.remove_entry(client_id, op_id, object); if (ret < 0) { cerr << "ERROR: failed to set state: " << cpp_strerror(-ret) << std::endl; return -ret; } } if (opt_cmd == OPT_REPLICALOG_GET || opt_cmd == OPT_REPLICALOG_UPDATE || opt_cmd == OPT_REPLICALOG_DELETE) { if (replica_log_type_str.empty()) { cerr << "ERROR: need to specify --replica-log-type=" << std::endl; return EINVAL; } } if (opt_cmd == OPT_REPLICALOG_GET) { RGWReplicaBounds bounds; if (replica_log_type == ReplicaLog_Metadata) { if (!specified_shard_id) { cerr << "ERROR: shard-id must be specified for get operation" << std::endl; return EINVAL; } RGWReplicaObjectLogger logger(store, pool, META_REPLICA_LOG_OBJ_PREFIX); int ret = logger.get_bounds(shard_id, bounds); if (ret < 0) return -ret; } else if (replica_log_type == ReplicaLog_Data) { if (!specified_shard_id) { cerr << "ERROR: shard-id must be specified for get operation" << std::endl; return EINVAL; } RGWReplicaObjectLogger logger(store, pool, DATA_REPLICA_LOG_OBJ_PREFIX); int ret = logger.get_bounds(shard_id, bounds); if (ret < 0) return -ret; } else if (replica_log_type == ReplicaLog_Bucket) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWReplicaBucketLogger logger(store); ret = logger.get_bounds(bucket, shard_id, bounds); if (ret < 0) return -ret; } else { // shouldn't get here ceph_abort(); } encode_json("bounds", bounds, formatter); formatter->flush(cout); cout << std::endl; } if (opt_cmd == OPT_REPLICALOG_DELETE) { if (replica_log_type == ReplicaLog_Metadata) { if (!specified_shard_id) { cerr << "ERROR: shard-id must be specified for delete operation" << std::endl; return EINVAL; } if (!specified_daemon_id) { cerr << "ERROR: daemon-id must be specified for delete operation" << std::endl; return EINVAL; } RGWReplicaObjectLogger logger(store, pool, META_REPLICA_LOG_OBJ_PREFIX); int ret = logger.delete_bound(shard_id, daemon_id, false); if (ret < 0) return -ret; } else if (replica_log_type == ReplicaLog_Data) { if (!specified_shard_id) { cerr << "ERROR: shard-id must be specified for delete operation" << std::endl; return EINVAL; } if (!specified_daemon_id) { cerr << "ERROR: daemon-id must be specified for delete operation" << std::endl; return EINVAL; } RGWReplicaObjectLogger logger(store, pool, DATA_REPLICA_LOG_OBJ_PREFIX); int ret = logger.delete_bound(shard_id, daemon_id, false); if (ret < 0) return -ret; } else if (replica_log_type == ReplicaLog_Bucket) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWReplicaBucketLogger logger(store); ret = logger.delete_bound(bucket, shard_id, daemon_id, false); if (ret < 0) return -ret; } } if (opt_cmd == OPT_REPLICALOG_UPDATE) { if (marker.empty()) { cerr << "ERROR: marker was not specified" < entries; int ret = read_decode_json(infile, entries); if (ret < 0) { cerr << "ERROR: failed to decode entries" << std::endl; return EINVAL; } RGWReplicaBounds bounds; if (replica_log_type == ReplicaLog_Metadata) { if (!specified_shard_id) { cerr << "ERROR: shard-id must be specified for get operation" << std::endl; return EINVAL; } RGWReplicaObjectLogger logger(store, pool, META_REPLICA_LOG_OBJ_PREFIX); int ret = logger.update_bound(shard_id, daemon_id, marker, time, &entries); if (ret < 0) { cerr << "ERROR: failed to update bounds: " << cpp_strerror(-ret) << std::endl; return -ret; } } else if (replica_log_type == ReplicaLog_Data) { if (!specified_shard_id) { cerr << "ERROR: shard-id must be specified for get operation" << std::endl; return EINVAL; } RGWReplicaObjectLogger logger(store, pool, DATA_REPLICA_LOG_OBJ_PREFIX); int ret = logger.update_bound(shard_id, daemon_id, marker, time, &entries); if (ret < 0) { cerr << "ERROR: failed to update bounds: " << cpp_strerror(-ret) << std::endl; return -ret; } } else if (replica_log_type == ReplicaLog_Bucket) { if (bucket_name.empty()) { cerr << "ERROR: bucket not specified" << std::endl; return EINVAL; } RGWBucketInfo bucket_info; int ret = init_bucket(tenant, bucket_name, bucket_id, bucket_info, bucket); if (ret < 0) { cerr << "ERROR: could not init bucket: " << cpp_strerror(-ret) << std::endl; return -ret; } RGWReplicaBucketLogger logger(store); ret = logger.update_bound(bucket, shard_id, daemon_id, marker, time, &entries); if (ret < 0) { cerr << "ERROR: failed to update bounds: " << cpp_strerror(-ret) << std::endl; return -ret; } } } bool quota_op = (opt_cmd == OPT_QUOTA_SET || opt_cmd == OPT_QUOTA_ENABLE || opt_cmd == OPT_QUOTA_DISABLE); if (quota_op) { if (bucket_name.empty() && user_id.empty()) { cerr << "ERROR: bucket name or uid is required for quota operation" << std::endl; return EINVAL; } if (!bucket_name.empty()) { if (!quota_scope.empty() && quota_scope != "bucket") { cerr << "ERROR: invalid quota scope specification." << std::endl; return EINVAL; } set_bucket_quota(store, opt_cmd, tenant, bucket_name, max_size, max_objects, have_max_size, have_max_objects); } else if (!user_id.empty()) { if (quota_scope == "bucket") { set_user_bucket_quota(opt_cmd, user, user_op, max_size, max_objects, have_max_size, have_max_objects); } else if (quota_scope == "user") { set_user_quota(opt_cmd, user, user_op, max_size, max_objects, have_max_size, have_max_objects); } else { cerr << "ERROR: invalid quota scope specification. Please specify either --quota-scope=bucket, or --quota-scope=user" << std::endl; return EINVAL; } } } return 0; }