Project

General

Profile

Actions

Tasks #47369

closed

Ceph scales to 100's of hosts, 1000's of OSDs....can orchestrator?

Added by Paul Cuzner over 3 years ago. Updated over 2 years ago.

Status:
Resolved
Priority:
Normal
Assignee:
-
Category:
orchestrator
Target version:
% Done:

0%

Tags:
Reviewed:
Affected Versions:
Pull request ID:

Description

This bug is intended to help us identify areas in mgr/orchestrator, mgr/cephadm, cephadm, mgr/prometheus and mgr/rook which represent potential scale bottlenecks.

For example
  • list_daemons
  • ceph_volume execution time
  • orchestrator task parallelization
  • lack of dashboard pagination
  • dashboard prefetch strategy for 100's or 1000s of OSDs, RBD images, buckets
  • mgr/prometheus reporting on 1000's OSDs or 100's hosts
  • installation bandwidth (the demand image pull places on the network)

There have been trackers in the past that focus on a specific areas (https://tracker.ceph.com/issues/36451), but it would be great if we could look at scale issues holistically..

Please feel free to add information to this tracker which documents scale issues with the management layer, that need to be considered.


Related issues 2 (1 open1 closed)

Related to Orchestrator - Feature #47368: Provide a daemon mode for cephadm to handle host/daemon state requestsResolvedPaul Cuzner

Actions
Related to Dashboard - Tasks #36451: mgr/dashboard: Scalability testingNewErnesto Puerta

Actions
Actions

Also available in: Atom PDF