Project

General

Profile

Bug #57506

Updated by Samuel Just over 1 year ago

<pre> 
   MDS=0 MGR=1 OSD=3 MON=1 ../src/vstart.sh --without-dashboard -X --crimson --redirect-output --debug -n --no-restart --crimson-smp 3 
   ./bin/ceph osd pool create rbd 32 32 replicated replicated_rule 2 2 2 
   ./bin/rados bench 1000 write --admin-socket asok/bench.asok -p rbd -b 4096 --debug-ms=1 2>out/bench.stderr 
   kill -9 688982 # kill osd.2 
   ./bin/ceph osd down 2 
   ./bin/ceph osd out 2 

 ./bin/ceph -s 

 *** DEVELOPER MODE: setting PATH, PYTHONPATH and LD_LIBRARY_PATH *** 
 2022-09-12T23:02:44.425+0000 7fa6c5a4c640 -1 WARNING: all dangerous and experimental features are enabled. 
 2022-09-12T23:02:44.428+0000 7fa6c5a4c640 -1 WARNING: all dangerous and experimental features are enabled. 
   cluster: 
     id:       07d892de-d0fb-4b86-b74d-cbc11d240a7a 
     health: HEALTH_WARN 
             Degraded data redundancy: 14449/44218 objects degraded (32.677%), 17 pgs degraded 
             1 pool(s) do not have an application enabled 
 
   services: 
     mon: 1 daemons, quorum a (age 7m) 
     mgr: x(active, since 7m) 
     osd: 3 osds: 2 up (since 62s), 2 in (since 21s); 17 remapped pgs 
 
   data: 
     pools:     2 pools, 33 pgs 
     objects: 22.11k objects, 86 MiB 
     usage:     2.2 GiB used, 200 GiB / 202 GiB avail 
     pgs:       14449/44218 objects degraded (32.677%) 
              15 active+clean 
              14 active+recovery_wait+undersized+degraded+remapped+wait 
              2    active+recovering+undersized+degraded+remapped+wait 
              1    active+undersized+wait 
              1    active+recovery_wait+undersized+degraded+remapped 
 
   io: 
     client:     122 KiB/s wr, 0 op/s rd, 30 op/s wr 
     recovery: 85 KiB/s, 21 objects/s 
 </pre> 

 Interestingly, IO continues anyway. 
 </pre>

Back