Tasks #17851
jewel v10.2.6
0%
Description
Workflow¶
- Preparing the release
- Cutting the release
- Abhishek V. asks Abhishek L. if a point release should be published - YES
- Abhishek V. gets approval from all leads
- Yehuda, rgw - YES (February 20, 2017 on ceph-devel mailing list)
- John, CephFS - YES (February 22, 2017 on ceph-devel mailing list)
- Jason, RBD - YES (February 20, 2017 on ceph-devel mailing list)
- Josh, rados - YES (February 21, 2017 on ceph-devel mailing list)
- Abhishek L. writes and commits the release notes
- Abhishek V. informs Yuri that the branch is ready for testing - DONE (February 22, 2017 on ceph-devel mailing list)
- Yuri runs additional integration tests - DONE (3/2/17 on ceph-devel mailing list)
- If Yuri discovers new bugs that need to be backported urgently (i.e. their priority is set to Urgent), the release goes back to being prepared, it was not ready after all
- Yuri informs Alfredo that the branch is ready for release - DONE
- Alfredo creates the packages and sets the release tag
Release information¶
- branch to build from: jewel, commit: d9eaab456ff45ae88e83bd633f0c4efb5902bf07
- version: v10.2.6
- type of release: point release
- where to publish the release: http://download.ceph.com/debian-jewel and http://download.ceph.com/rpm-jewel
History
#1 Updated by Loïc Dachary over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel-next..jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 10865
- + Pull request 11413
- + Pull request 11470
- |\
- | + rgw:bucket check remove multipart prefix
- + Pull request 11476
- |\
- | + rgw: RGWCoroutinesManager::run returns status of last cr
- + Pull request 11477
- |\
- | + rgw: fix for assertion in RGWMetaSyncCR
- + Pull request 11508
- |\
- | + utime.h: fix timezone issue in round_to_* funcs.
- + Pull request 11529
- |\
- | + common: Improve linux dcache hash algorithm
- + Pull request 11574
- + Pull request 11606
- + Pull request 11627
- + Pull request 11660
- |\
- | + mon/PGMap: PGs can be stuck more than one thing
- + Pull request 11672
- |\
- | + rgw_rest_s3: apply missed base64 try-catch
- + Pull request 11675
- + Pull request 11735
- + Pull request 11736
- + Pull request 11737
- + Pull request 11743
- + Pull request 11757
- |\
- | + rgw ldap: protect rgw::from_base64 from non-base64 input
- + Pull request 11758
- |\
- | + rgw: fix osd crashes when execute radosgw-admin bi list --max-entries=1 command
- + Pull request 11759
- |\
- | + rgw: json encode/decode index_type, allow modification
- + Pull request 11760
- + Pull request 11852
- + Pull request 11853
- + Pull request 11854
- + Pull request 11855
- |\
- | + rpm: fix permissions for /etc/ceph/rbdmap
- + Pull request 11856
- + Pull request 11857
- + Pull request 11858
- + Pull request 11860
- + Pull request 11861
- + Pull request 11862
- + Pull request 11863
- + Pull request 11864
- + Pull request 11865
- |\
- | + rgw: RGWSimpleRadosReadCR tolerates empty reads
- + Pull request 11866
- |\
- | + rgw: clean up RGWShardedOmapCRManager on early return
- + Pull request 11867
- + Pull request 11868
- |\
- | + rgw: store oldest mdlog period in rados
- + Pull request 11869
- + Pull request 11870
- + Pull request 11871
- |\
- | + librbd: batch ObjectMap updations upon trim
- + Pull request 11872
- |\
- | + rgw: delete entries_index in RGWFetchAllMetaCR
- + Pull request 11873
- + Pull request 11875
- + Pull request 11876
- |\
- | + rgw: fix the field 'total_time' of log entry in log show opt
- + Pull request 11884
- + ceph-create-keys: wait 10 minutes to get or create the bootstrap key, not forever
- + ceph-create-keys: wait 10 minutes to get or create a key, not forever
- + ceph-create-keys: wait for quorum for ten minutes, not forever
#2 Updated by Loïc Dachary over 6 years ago
rbd¶
teuthology-suite --priority 1000 --suite rbd --subset $(expr $RANDOM % 5)/5 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2016-11-10_10:23:37-rbd-jewel-backports---basic-smithi/
- 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=b124ab8b49b6708673caf6b240d73515122f8eb3 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror.sh'
- 'mkdir
Re-running failed tests
#3 Updated by Loïc Dachary over 6 years ago
rgw¶
teuthology-suite -k distro --priority 1000 --suite rgw --subset $(expr $RANDOM % 5)/5 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2016-11-10_10:25:55-rgw-jewel-backports-distro-basic-smithi
- HTTPConnectionPool(host='smithi075.front.sepia.ceph.com', port=8000): Max retries exceeded with url: /metadata/incremental (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x7fb10a323dd0>: Failed to establish a new connection: [Errno 111] Connection refused',))
Re-running failed tests
#4 Updated by Loïc Dachary over 6 years ago
rados¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 2000)/2000 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2016-11-10_10:27:02-rados-jewel-backports-distro-basic-smithi
- 'cd /home/ubuntu/cephtest && sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term valgrind --trace-children=no --child-silent-after-fork=yes --num-callers=50 --suppressions=/home/ubuntu/cephtest/valgrind.supp --xml=yes --xml-file=/var/log/ceph/valgrind/mon.b.log --time-stamp=yes --tool=memcheck --leak-check=full --show-reachable=yes ceph-mon -f --cluster ceph -i b'
- 'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
Re-running failed tests
#5 Updated by Loïc Dachary over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --subset $(expr $RANDOM % 5)/5 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2016-11-10_10:29:25-fs-jewel-backports-distro-basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=b124ab8b49b6708673caf6b240d73515122f8eb3 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/libcephfs/test.sh'- fs/basic/{clusters/fixed-2-ucephfs.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml fs/btrfs.yaml inline/no.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml}
- fs/verify/{clusters/fixed-2-ucephfs.yaml debug/{mds_client.yaml mon.yaml} dirfrag/frag_enable.yaml fs/btrfs.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml validater/valgrind.yaml}
- fs/basic/{clusters/fixed-2-ucephfs.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml fs/btrfs.yaml inline/yes.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml}
- fs/verify/{clusters/fixed-2-ucephfs.yaml debug/{mds_client.yaml mon.yaml} dirfrag/frag_enable.yaml fs/btrfs.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml validater/lockdep.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=b124ab8b49b6708673caf6b240d73515122f8eb3 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/libcephfs-java/test.sh' - 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
- 'mkdir
Re-running failed tests
- fail http://pulpito.ceph.com/loic-2016-11-11_10:57:34-fs-jewel-backports-distro-basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=b124ab8b49b6708673caf6b240d73515122f8eb3 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/libcephfs/test.sh'- fs/basic/{clusters/fixed-2-ucephfs.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml fs/btrfs.yaml inline/no.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml}
- fs/verify/{clusters/fixed-2-ucephfs.yaml debug/{mds_client.yaml mon.yaml} dirfrag/frag_enable.yaml fs/btrfs.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml validater/valgrind.yaml}
- fs/basic/{clusters/fixed-2-ucephfs.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml fs/btrfs.yaml inline/yes.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml}
- fs/verify/{clusters/fixed-2-ucephfs.yaml debug/{mds_client.yaml mon.yaml} dirfrag/frag_enable.yaml fs/btrfs.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml validater/lockdep.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=b124ab8b49b6708673caf6b240d73515122f8eb3 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/libcephfs-java/test.sh'
- 'mkdir
#6 Updated by Loïc Dachary over 6 years ago
powercycle¶
teuthology-suite -v -c jewel-backports --suite-branch jewel -k testing -m smithi -s powercycle -p 1000 --email loic@dachary.org
#7 Updated by Loïc Dachary over 6 years ago
Upgrade¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x --suite-branch jewel --ceph jewel-backports --machine-type vps --priority 1000 machine_types/vps.yaml
- fail http://pulpito.ceph.com/loic-2016-11-10_10:31:33-upgrade:jewel-x-jewel-backports-distro-basic-vps
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=jewel TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh'
- 'mkdir
Re-running failed tests
- pass http://pulpito.ceph.com/loic-2016-11-11_10:55:56-upgrade:jewel-x-jewel-backports-distro-basic-vps
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --suite-branch jewel --ceph jewel-backports --machine-type vps --priority 1000 machine_types/vps.yaml
- fail http://pulpito.ceph.com/loic-2016-11-10_10:32:38-upgrade:hammer-x-jewel-backports-distro-basic-vps
- 'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
Re-running failed tests
- fail http://pulpito.ceph.com/loic-2016-11-11_10:52:59-upgrade:hammer-x-jewel-backports-distro-basic-vps
Re-running the dead test with https://github.com/ceph/ceph-qa-suite/pull/1256 to set require_jewel
filter='upgrade:hammer-x/stress-split/{0-tz-eastern.yaml 0-cluster/{openstack.yaml start.yaml} 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_14.04.yaml}' teuthology-suite -k distro --filter="$filter" --verbose --suite upgrade/hammer-x --suite-branch wip-17734-jewel --ceph jewel-backports --machine-type vps --priority 1000 machine_types/vps.yaml
#8 Updated by Loïc Dachary over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --suite-branch jewel --ceph jewel-backports --machine-type vps --priority 1000 machine_types/vps.yaml
#9 Updated by Loïc Dachary over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel-next..jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 11470
- |\
- | + rgw:bucket check remove multipart prefix
- + Pull request 11476
- |\
- | + rgw: RGWCoroutinesManager::run returns status of last cr
- + Pull request 11477
- |\
- | + rgw: fix for assertion in RGWMetaSyncCR
- + Pull request 11497
- |\
- | + rgw: settle /info implementation across other swift-at-root features.
- | + swift /info implementation.
- | + rgw: add support for the healthcheck feature of Swift API.
- | + rgw: add support for the crossdomain.xml resource of Swift API.
- | + rgw: fix the handling of rgw_swift_url_prefix.
- | + rgw_http_errors: add http error code for 503
- | + rgw: Allow to serve Swift off the URL root
- + Pull request 11508
- |\
- | + utime.h: fix timezone issue in round_to_* funcs.
- + Pull request 11529
- |\
- | + common: Improve linux dcache hash algorithm
- + Pull request 11656
- |\
- | + ceph_volume_client: fix partial auth recovery
- | + ceph_volume_client: check if volume metadata is empty
- | + ceph_volume_client: fix _recover_auth_meta() method
- + Pull request 11672
- |\
- | + rgw_rest_s3: apply missed base64 try-catch
- + Pull request 11758
- |\
- | + rgw: fix osd crashes when execute radosgw-admin bi list --max-entries=1 command
- + Pull request 11759
- |\
- | + rgw: json encode/decode index_type, allow modification
- + Pull request 11865
- |\
- | + rgw: RGWSimpleRadosReadCR tolerates empty reads
- + Pull request 11866
- |\
- | + rgw: clean up RGWShardedOmapCRManager on early return
- + Pull request 11868
- |\
- | + rgw: store oldest mdlog period in rados
- + Pull request 11872
- |\
- | + rgw: delete entries_index in RGWFetchAllMetaCR
- + Pull request 11876
- |\
- | + rgw: fix the field 'total_time' of log entry in log show opt
- + Pull request 11884
- |\
- | + ceph-create-keys: wait 10 minutes to get or create the bootstrap key, not forever
- | + ceph-create-keys: wait 10 minutes to get or create a key, not forever
- | + ceph-create-keys: wait for quorum for ten minutes, not forever
- + Pull request 11944
- |\
- | + osd: Add config option to disable new scrubs during recovery
- + Pull request 11947
- |\
- | + mon: update mon(peon)'s down_pending_out when osd up
- + Pull request 11953
- |\
- | + test: temporarily disable fork()'ing tests
- + Pull request 11968
- |\
- | + qa/workunits: update test_cache_pool.sh
- | + tools/rados: add --with-clones option to include clones for cache-flush/cache-evict
- | + tools/rados: default to include clone objects when excuting cache-flush-evict-all
- + Pull request 11990
- |\
- | + rgw: for the create_bucket api, if the input creation_time is zero, we should set it to 'now
- + Pull request 11991
- |\
- | + osd: limit omap data in push op
- + Pull request 11997
- |\
- | + ReplicatedPG::do_update_log_missing: take the pg lock in the callback
- + Pull request 11998
- |\
- | + test: add test for fiemap xfs issue when #extents > 1364
- | + FileStore:: fix fiemap issue in xfs when #extents > 1364
- + Pull request 11999
- |\
- | + mon: MonmapMonitor: drop unnecessary 'goto' statements
- | + mon: MonmapMonitor: return success when monitor will be removed
- + Pull request 12001
- |\
- | + os/filestore/HashIndex: fix list_by_hash_* termination on reaching end
- + Pull request 12033
- |\
- | + mon,ceph-disk: add lockbox permissions to bootstrap-osd
- + Pull request 12043
- |\
- | + rbd-mirror: Add sparse read for sync image
- + Pull request 12044
- |\
- | + rgw: fix for bucket delete racing with mdlog sync
- + Pull request 12045
- |\
- | + rgw_file: fix spurious mount entries w/Linux NFS client
- + Pull request 12067
- |\
- | + OSDMonitor: only reject MOSDBoot based on up_from if inst matches
- + Pull request 12079
- |\
- | + rgw: TempURL properly handles accounts created with the implicit tenant.
- + Pull request 12088
- |\
- | + test.sh: Make check for flags more robust
- | + test: Remove extra objectstore_tool call which causes a recovery
- | + test: Handle object removals in a non-racey way
- | + osd: Fix hang on unfound object after mark_unfound_lost is done
- | + osd: Handle recovery read errors
- | + osd: Fix log messages
- | + osd: CLEANUP: Remove unused pending_read member
- | + Revert test: Disable tests due to recovery race
- | + test/osd-scrub-repair.sh: Use test case specific object names to help with diagnostics
- + Pull request 12137
- |\
- | + client: fix stale entries in command table
- + Pull request 12147
- |\
- | + ceph-disk: trigger must ensure device ownership
- | + ceph-disk: systemd unit must run after local-fs.target
- + Pull request 12151
- |\
- | + tests: save 9 characters for asok paths
- + Pull request 12153
- |\
- | + mds: ignore 'session evict' when mds is replaying log
- + Pull request 12154
- |\
- | + mds: use projected path construction for access
- + Pull request 12155
- |\
- | + mds: require MAY_SET_POOL to set pool_ns
- + Pull request 12156
- + rgw: look for region_map in rgw_region_root_pool
- + rgw: region conversion respects pre-existing rgw_region_root_pool
#10 Updated by Loïc Dachary over 6 years ago
rbd¶
teuthology-suite --priority 1000 --suite rbd --subset $(expr $RANDOM % 5)/5 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/loic-2016-11-23_21:22:04-rbd-jewel-backports---basic-smithi/
- 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4920d41d4d8be57c06687d4de029ec2b687a9f09 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror_stress.sh' - 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4920d41d4d8be57c06687d4de029ec2b687a9f09 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror.sh'- rbd/mirror/{base/install.yaml cluster/{2-node.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml rbd-mirror/one-per-cluster.yaml workloads/rbd-mirror-workunit.yaml}
- rbd/mirror/{base/install.yaml cluster/{2-node.yaml openstack.yaml} fs/xfs.yaml msgr-failures/many.yaml rbd-mirror/one-per-cluster.yaml workloads/rbd-mirror-workunit.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4920d41d4d8be57c06687d4de029ec2b687a9f09 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin IMAGE_NAME=client.0.1-clone adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/qemu_rebuild_object_map.sh'
- 'mkdir
Re-running the failed jobs
- fail http://pulpito.ceph.com/loic-2016-11-25_08:54:08-rbd-jewel-backports---basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4920d41d4d8be57c06687d4de029ec2b687a9f09 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror_stress.sh' - 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term /usr/libexec/qemu-kvm -enable-kvm -nographic -m 4096 -drive file=/home/ubuntu/cephtest/qemu/base.client.0.qcow2,format=qcow2,if=virtio -cdrom /home/ubuntu/cephtest/qemu/client.0.iso -drive file=rbd:rbd/client.0.0-clone:id=0,format=raw,if=virtio,cache=none -drive file=rbd:rbd/client.0.1-clone:id=0,format=raw,if=virtio,cache=none'
- 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4920d41d4d8be57c06687d4de029ec2b687a9f09 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror.sh'
- 'mkdir
Re-running failed jobs
Re-running failed jobs on jewel
#11 Updated by Loïc Dachary over 6 years ago
rgw¶
teuthology-suite -k distro --priority 1000 --suite rgw --subset $(expr $RANDOM % 5)/5 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/loic-2016-11-23_21:26:42-rgw-jewel-backports-distro-basic-smithi/
- **
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/ec-profile.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/ec-cache.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/replicated.yaml}
- **
Re-running failed jobs
#12 Updated by Loïc Dachary over 6 years ago
rados¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 2000)/2000 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
#13 Updated by Loïc Dachary over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --subset $(expr $RANDOM % 5)/5 --suite-branch jewel --email loic@dachary.org --ceph jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/loic-2016-11-23_21:37:30-fs-jewel-backports-distro-basic-smithi/
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4920d41d4d8be57c06687d4de029ec2b687a9f09 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/libcephfs-java/test.sh' - 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
- 'mkdir
Re-running failed jobs
- fail http://pulpito.front.sepia.ceph.com/loic-2016-11-25_08:58:04-fs-jewel-backports-distro-basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=4920d41d4d8be57c06687d4de029ec2b687a9f09 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/libcephfs-java/test.sh'
- 'mkdir
#14 Updated by Loïc Dachary over 6 years ago
powercycle¶
teuthology-suite -v -c jewel-backports --suite-branch jewel -k distro -m smithi -s powercycle -p 1000 --email loic@dachary.org
- fail http://pulpito.ceph.com:80/loic-2016-11-23_21:39:07-powercycle-jewel-backports-distro-basic-smithi/
- 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 2'
Re-running the failed job
#15 Updated by Loïc Dachary over 6 years ago
Upgrade¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x --suite-branch jewel --ceph jewel-backports --machine-type vps --priority 1000 machine_types/vps.yaml
- fail http://pulpito.ceph.com/loic-2016-11-23_21:39:57-upgrade:jewel-x-jewel-backports-distro-basic-vps
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=jewel TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh'
- 'mkdir
Re-running failed jobs
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --suite-branch jewel --ceph jewel-backports --machine-type vps --priority 1000 machine_types/vps.yaml
- fail http://pulpito.ceph.com/loic-2016-11-23_21:46:45-upgrade:hammer-x-jewel-backports-distro-basic-vps
- File is closed
- upgrade:hammer-x/stress-split-erasure-code/{0-tz-eastern.yaml 0-cluster/{openstack.yaml start.yaml} 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/ubuntu_14.04.yaml}
- upgrade:hammer-x/stress-split-erasure-code-x86_64/{0-tz-eastern.yaml 0-x86_64.yaml 0-cluster/{openstack.yaml start.yaml} 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-rados-plugin=isa-k=2-m=1.yaml}
- 'sudo TESTDIR=/home/ubuntu/cephtest bash -c \'sudo ceph osd erasure-code-profile set profile-shec k=2 m=1 c=1 plugin=shec 2>&1 | grep "unsupported by"\''
- upgrade:hammer-x/stress-split-erasure-code/{0-tz-eastern.yaml 0-cluster/{openstack.yaml start.yaml} 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-no-shec.yaml distros/ubuntu_14.04.yaml}
- upgrade:hammer-x/stress-split-erasure-code/{0-tz-eastern.yaml 0-cluster/{openstack.yaml start.yaml} 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-no-shec.yaml distros/centos_7.2.yaml}
- File is closed
Re-running failed jobs
#16 Updated by Loïc Dachary over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --suite-branch jewel --ceph jewel-backports --machine-type vps --priority 1000 machine_types/vps.yaml
#17 Updated by Loïc Dachary over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel-next..wip-jewel-backports-loic | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 10661
- |\
- | + rgw: Have a flavor of bucket deletion to bypass GC and to trigger object deletions async.
- + Pull request 11470
- |\
- | + rgw:bucket check remove multipart prefix
- + Pull request 11476
- |\
- | + rgw: RGWCoroutinesManager::run returns status of last cr
- + Pull request 11477
- |\
- | + rgw: fix for assertion in RGWMetaSyncCR
- + Pull request 11497
- |\
- | + rgw: settle /info implementation across other swift-at-root features.
- | + swift /info implementation.
- | + rgw: add support for the healthcheck feature of Swift API.
- | + rgw: add support for the crossdomain.xml resource of Swift API.
- | + rgw: fix the handling of rgw_swift_url_prefix.
- | + rgw_http_errors: add http error code for 503
- | + rgw: Allow to serve Swift off the URL root
- + Pull request 11508
- |\
- | + utime.h: fix timezone issue in round_to_* funcs.
- + Pull request 11672
- |\
- | + rgw_rest_s3: apply missed base64 try-catch
- + Pull request 11759
- |\
- | + rgw: json encode/decode index_type, allow modification
- + Pull request 11865
- |\
- | + rgw: RGWSimpleRadosReadCR tolerates empty reads
- + Pull request 11866
- |\
- | + rgw: clean up RGWShardedOmapCRManager on early return
- + Pull request 11868
- |\
- | + rgw: store oldest mdlog period in rados
- + Pull request 11872
- |\
- | + rgw: delete entries_index in RGWFetchAllMetaCR
- + Pull request 11876
- |\
- | + rgw: fix the field 'total_time' of log entry in log show opt
- + Pull request 11944
- |\
- | + osd: Add config option to disable new scrubs during recovery
- + Pull request 11947
- |\
- | + mon: update mon(peon)'s down_pending_out when osd up
- + Pull request 11953
- |\
- | + test: temporarily disable fork()'ing tests
- + Pull request 11968
- |\
- | + qa/workunits: update test_cache_pool.sh
- | + tools/rados: add --with-clones option to include clones for cache-flush/cache-evict
- | + tools/rados: default to include clone objects when excuting cache-flush-evict-all
- + Pull request 11990
- |\
- | + rgw: for the create_bucket api, if the input creation_time is zero, we should set it to 'now
- + Pull request 11991
- |\
- | + osd: limit omap data in push op
- + Pull request 11997
- |\
- | + ReplicatedPG::do_update_log_missing: take the pg lock in the callback
- + Pull request 12033
- |\
- | + mon,ceph-disk: add lockbox permissions to bootstrap-osd
- + Pull request 12043
- |\
- | + rbd-mirror: Add sparse read for sync image
- + Pull request 12044
- |\
- | + rgw: fix for bucket delete racing with mdlog sync
- + Pull request 12045
- |\
- | + rgw_file: fix spurious mount entries w/Linux NFS client
- + Pull request 12079
- |\
- | + rgw: TempURL properly handles accounts created with the implicit tenant.
- + Pull request 12088
- |\
- | + test.sh: Make check for flags more robust
- | + test: Remove extra objectstore_tool call which causes a recovery
- | + test: Handle object removals in a non-racey way
- | + osd: Fix hang on unfound object after mark_unfound_lost is done
- | + osd: Handle recovery read errors
- | + osd: Fix log messages
- | + osd: CLEANUP: Remove unused pending_read member
- | + test/osd-scrub-repair.sh: Use test case specific object names to help with diagnostics
- + Pull request 12137
- |\
- | + client: fix stale entries in command table
- + Pull request 12147
- |\
- | + ceph-disk: enable --runtime ceph-osd systemd units
- | + build/ops: restart ceph-osd@.service after 20s instead of 100ms
- | + ceph-disk: trigger must ensure device ownership
- | + ceph-disk: systemd unit must run after local-fs.target
- + Pull request 12151
- |\
- | + tests: save 9 characters for asok paths
- + Pull request 12153
- |\
- | + mds: ignore 'session evict' when mds is replaying log
- + Pull request 12154
- |\
- | + mds: use projected path construction for access
- + Pull request 12155
- |\
- | + mds: require MAY_SET_POOL to set pool_ns
- + Pull request 12156
- |\
- | + rgw: look for region_map in rgw_region_root_pool
- | + rgw: region conversion respects pre-existing rgw_region_root_pool
- + Pull request 12159
- |\
- | + qa/workunits/rbd: check status also in pool dir after asok commands
- | + qa/workunits/rbd: wait for image deleted before checking health
- | + qa/workunits/rbd: wait for image deleted before checking health
- | + qa/workunits/rbd: small fixup and improvements for rbd-mirror tests
- + Pull request 12210
- |\
- | + systemd/ceph-disk: reduce ceph-disk flock contention
- + Pull request 12239
- |\
- | + jewel: fixed the issue when --disable-server, compilation fails.
- + Pull request 12258
- |\
- | + rgw: add support for the prefix parameter in account listing of Swift API.
- | + rgw: optimize out ctor-copy in RGWListBuckets_ObjStore_SWIFT.
- + Pull request 12291
- |\
- | + msg/simple/Pipe: handle addr decode error
- + Pull request 12296
- |\
- | + build/ops: fix undefined crypto references with --with-xio
- + Pull request 12302
- |\
- | + tests: check hostname --fqdn sanity before running cmake check
- | + tests: check hostname --fqdn sanity before running make check
- + Pull request 12313
- |\
- | + rgw: only set CURLOPT_UPLOAD for PUT/POST requests
- + Pull request 12314
- |\
- | + rgw: RGWBucketSyncStatusManager uses existing async_rados
- + Pull request 12315
- |\
- | + rgw: fix missing master zone for a single zone zonegroup
- + Pull request 12316
- |\
- | + rgw: add recovery procedure for upgrade to older version of jewel
- + Pull request 12320
- |\
- | + rgw: fix for versioned delete_multi_object
- | + rgw:fix for deleting objects name beginning and ending with underscores of one bucket using POST method of AWS's js sdk. Fixes: http://tracker.ceph.com/issues/17888
- + Pull request 12321
- |\
- | + qa/workunits/rbd: use image id when probing for image presence
- + Pull request 12322
- |\
- | + librbd: diffs to clone's first snapshot should include parent diffs
- + Pull request 12323
- |\
- | + librbd: account m_processing when failing request after refresh
- + Pull request 12324
- |\
- | + mds: force client flush snap data before truncating objects
- + Pull request 12325
- |\
- | + ceph_volume_client: set an existing auth ID's default mon caps
- + Merge remote-tracking branch 'ceph/jewel-next' into wip-jewel-backports-loic
- + Pull request 12167
- |\
- | + crush: condition latest tunable encoding on features
- | + crush/CrushWrapper: encode with features
- | + crush/CrushWrapper: drop unused 'lean' encode() argument
- | + osd/osd_types: encode pg_pool_t like hammer if features indicate hammer
- | + osd/osd_types: conditional pg_pool_t encoding
- + Pull request 12067
- |\
- | + OSDMonitor: only reject MOSDBoot based on up_from if inst matches
- |/
- + Pull request 12207
- |\
- | + librados: remove new setxattr overload to avoid breaking the C++ ABI
- + Pull request 12267
- |\
- | + mon: MonmapMonitor: drop unnecessary 'goto' statements
- | + mon: MonmapMonitor: return success when monitor will be removed
- |/
- + Pull request 12001
- + os/filestore/HashIndex: fix list_by_hash_* termination on reaching end
#18 Updated by Loïc Dachary over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --suite-branch jewel --ceph wip-jewel-backports-loic --machine-type vps --priority 1000 machine_types/vps.yaml ~/shaman.yaml
#19 Updated by Loïc Dachary over 6 years ago
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --suite-branch jewel --ceph wip-jewel-backports-loic --machine-type vps --priority 1000 machine_types/vps.yaml ~/shaman.yaml
- fail http://pulpito.ceph.com/loic-2016-12-06_09:38:33-upgrade:hammer-x-wip-jewel-backports-loic-distro-basic-vps
- ceph-objectstore-tool: exp list-pgs failure with status 1
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=95cefea9fd9ab740263bf8bb4796fd864d9afe2b
- 'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=centos%2F7%2Fx86_64&ref=firefly
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&ref=firefly
#20 Updated by Loïc Dachary over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports-loic --suite-branch jewel -k distro -m smithi -s powercycle -p 1000 --email loic@dachary.org ~/shaman.yaml
- fail http://pulpito.ceph.com/loic-2016-12-06_09:40:31-powercycle-wip-jewel-backports-loic-distro-basic-smithi
- , 'smithi057.front.sepia.ceph.com': {'msg': 'One or more items failed', 'failed': True, 'changed': False}}
- 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 0'
#21 Updated by Loïc Dachary over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports-loic --machine-type smithi ~/shaman.yaml
- fail http://pulpito.ceph.com/loic-2016-12-06_09:42:04-fs-wip-jewel-backports-loic-distro-basic-smithi
- }
- saw valgrind issues
- 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
#22 Updated by Loïc Dachary over 6 years ago
rados¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 50)/50 --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports-loic --machine-type smithi ~/shaman.yaml
- fail http://pulpito.ceph.com/loic-2016-12-06_09:43:31-rados-wip-jewel-backports-loic-distro-basic-smithi
- SELinux denials found on ubuntu@smithi013.front.sepia.ceph.com: ['type=AVC msg=audit(1481196479.614:8253): avc: denied { read } for pid=10410 comm="ceph-osd" name="type" dev="nvme0n1p3" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481196520.457:8326): avc: denied { read } for pid=11590 comm="ceph-osd" name="type" dev="nvme0n1p3" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481196479.614:8253): avc: denied { open } for pid=10410 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-1/type" dev="nvme0n1p3" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481196520.457:8326): avc: denied { open } for pid=11590 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-1/type" dev="nvme0n1p3" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file']
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&ref=v0.80.8
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=9d446bd416c52cd785ccf048ca67737ceafcdd7f
- 'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
- saw valgrind issues
- rados/verify/{rados.yaml 1thrash/default.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr/async.yaml msgr-failures/few.yaml tasks/rados_api_tests.yaml validater/valgrind.yaml}
- rados/verify/{rados.yaml 1thrash/default.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr/simple.yaml msgr-failures/few.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml}
#23 Updated by Loïc Dachary over 6 years ago
rgw¶
teuthology-suite -k distro --priority 1000 --suite rgw --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports-loic --machine-type smithi ~/shaman.yaml
- fail http://pulpito.ceph.com/loic-2016-12-06_09:46:40-rgw-wip-jewel-backports-loic-distro-basic-smithi
- **
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/replicated.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/ec-cache.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/replicated.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/ec.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/ec.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/ec-profile.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/ec-profile.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/ec-cache.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/replicated.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/ec-cache.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/replicated.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/ec.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/ec.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/ec-profile.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-multi-region.yaml frontend/civetweb.yaml fs/xfs.yaml rgw_pool_type/ec-profile.yaml}
- rgw/singleton/{overrides.yaml xfs.yaml all/radosgw-admin-data-sync.yaml frontend/apache.yaml fs/xfs.yaml rgw_pool_type/ec-cache.yaml}
- HTTPConnectionPool(host='smithi126.front.sepia.ceph.com', port=8000): Max retries exceeded with url: /metadata/incremental (Caused by NewConnectionError('<requests.packages.urllib3.connection.HTTPConnection object at 0x7fa208f19e10>: Failed to establish a new connection: [Errno 111] Connection refused',))
- "SWIFT_TEST_CONFIG_FILE=/home/ubuntu/cephtest/archive/testswift.client.0.conf /home/ubuntu/cephtest/swift/virtualenv/bin/nosetests -w /home/ubuntu/cephtest/swift/test/functional -v -a '!fails_on_rgw'"
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_swift.yaml validater/lockdep.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec.yaml tasks/rgw_swift.yaml validater/lockdep.yaml}
- '/home/ubuntu/cephtest/s3-tests/virtualenv/bin/s3tests-test-readwrite'
- saw valgrind issues
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/replicated.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/replicated.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/replicated.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-profile.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/replicated.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-profile.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/replicated.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{overrides.yaml clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml rgw_pool_type/ec-profile.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- **
#24 Updated by Loïc Dachary over 6 years ago
rbd¶
teuthology-suite --priority 1000 --suite rbd --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports-loic --machine-type smithi ~/shaman.yaml
- fail http://pulpito.ceph.com/loic-2016-12-06_09:46:56-rbd-wip-jewel-backports-loic---basic-smithi
- 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term /usr/libexec/qemu-kvm -enable-kvm -nographic -m 4096 -drive file=/home/ubuntu/cephtest/qemu/base.client.0.qcow2,format=qcow2,if=virtio -cdrom /home/ubuntu/cephtest/qemu/client.0.iso -drive file=rbd:rbd/client.0.0-clone:id=0,format=raw,if=virtio,cache=writeback -drive file=rbd:rbd/client.0.1-clone:id=0,format=raw,if=virtio,cache=writeback'
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=10328195aad09f59d5c2c382bd9241c7418f744e TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/qemu-iotests.sh' - 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term /usr/libexec/qemu-kvm -enable-kvm -nographic -m 4096 -drive file=/home/ubuntu/cephtest/qemu/base.client.0.qcow2,format=qcow2,if=virtio -cdrom /home/ubuntu/cephtest/qemu/client.0.iso -drive file=rbd:rbd/client.0.0-clone:id=0,format=raw,if=virtio,cache=writethrough -drive file=rbd:rbd/client.0.1-clone:id=0,format=raw,if=virtio,cache=writethrough'
- SELinux denials found on ubuntu@smithi023.front.sepia.ceph.com: ['type=AVC msg=audit(1481252822.592:36797): avc: denied { create } for pid=13771 comm="mandb" name="13771" scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.561:36792): avc: denied { unlink } for pid=13771 comm="mandb" name="index.db" dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.561:36792): avc: denied { remove_name } for pid=13771 comm="mandb" name="#index.db#" dev="sda1" ino=29363155 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=dir', 'type=AVC msg=audit(1481252822.462:36790): avc: denied { create } for pid=13771 comm="mandb" name="#index.db#" scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252821.938:36784): avc: denied { create } for pid=13760 comm="logrotate" name="logrotate.status.tmp" scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252821.939:36785): avc: denied { setattr } for pid=13760 comm="logrotate" name="logrotate.status.tmp" dev="sda1" ino=29363169 scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.561:36792): avc: denied { rename } for pid=13771 comm="mandb" name="#index.db#" dev="sda1" ino=29363155 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.182:36789): avc: denied { lock } for pid=13771 comm="mandb" path="/var/cache/man/index.db" dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.561:36792): avc: denied { add_name } for pid=13771 comm="mandb" name="index.db" dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=dir', 'type=AVC msg=audit(1481252821.938:36784): avc: denied { write } for pid=13760 comm="logrotate" path="/var/lib/logrotate.status.tmp" dev="sda1" ino=29363169 scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.182:36788): avc: denied { open } for pid=13771 comm="mandb" path="/var/cache/man/index.db" dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.026:36786): avc: denied { rename } for pid=13760 comm="logrotate" name="logrotate.status.tmp" dev="sda1" ino=29363169 scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.462:36790): avc: denied { add_name } for pid=13771 comm="mandb" name="#index.db#" scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=dir', 'type=AVC msg=audit(1481252822.562:36795): avc: denied { read write } for pid=13771 comm="mandb" path="/var/cache/man/index.db" dev="sda1" ino=29363155 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.561:36792): avc: denied { write } for pid=13771 comm="mandb" name="man" dev="sda1" ino=29360274 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=dir', 'type=AVC msg=audit(1481252821.786:36783): avc: denied { read } for pid=13760 comm="logrotate" name="logrotate.status" dev="sda1" ino=29363155 scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.561:36793): avc: denied { lock } for pid=13771 comm="mandb" path=2F7661722F63616368652F6D616E2F23696E6465782E646223202864656C6574656429 dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.562:36794): avc: denied { getattr } for pid=13771 comm="mandb" path="/var/cache/man/index.db" dev="sda1" ino=29363155 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252821.786:36783): avc: denied { open } for pid=13760 comm="logrotate" path="/var/lib/logrotate.status" dev="sda1" ino=29363155 scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.182:36788): avc: denied { read write } for pid=13771 comm="mandb" name="index.db" dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252821.786:36782): avc: denied { getattr } for pid=13760 comm="logrotate" path="/var/lib/logrotate.status" dev="sda1" ino=29363155 scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.595:36798): avc: denied { setattr } for pid=13771 comm="mandb" name="13771" dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.026:36786): avc: denied { unlink } for pid=13760 comm="logrotate" name="logrotate.status" dev="sda1" ino=29363155 scontext=system_u:system_r:logrotate_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252823.146:36812): avc: denied { read } for pid=13771 comm="mandb" name="man" dev="sda1" ino=29360274 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=dir', 'type=AVC msg=audit(1481252822.157:36787): avc: denied { getattr } for pid=13771 comm="mandb" path="/var/cache/man/index.db" dev="sda1" ino=29361527 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481252822.462:36790): avc: denied { write } for pid=13771 comm="mandb" name="man" dev="sda1" ino=29360274 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=dir', 'type=AVC msg=audit(1481252822.592:36796): avc: denied { open } for pid=13771 comm="mandb" path="/var/cache/man/index.db" dev="sda1" ino=29363155 scontext=system_u:system_r:mandb_t:s0-s0:c0.c1023 tcontext=system_u:object_r:unlabeled_t:s0 tclass=file']
- }, 'changed': False, 'msg': 'Failed to update apt cache.'}}
- 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term /usr/libexec/qemu-kvm -enable-kvm -nographic -m 4096 -drive file=/home/ubuntu/cephtest/qemu/base.client.0.qcow2,format=qcow2,if=virtio -cdrom /home/ubuntu/cephtest/qemu/client.0.iso -drive file=rbd:rbd/client.0.0-clone:id=0,format=raw,if=virtio,cache=none -drive file=rbd:rbd/client.0.1-clone:id=0,format=raw,if=virtio,cache=none'
- rbd/qemu/{cache/none.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} features/defaults.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/qemu_xfstests.yaml}
- rbd/maintenance/{xfs.yaml base/install.yaml clusters/{fixed-3.yaml openstack.yaml} qemu/xfstests.yaml workloads/dynamic_features.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=10328195aad09f59d5c2c382bd9241c7418f744e TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror_stress.sh'- rbd/mirror/{base/install.yaml cluster/{2-node.yaml openstack.yaml} fs/xfs.yaml msgr-failures/many.yaml rbd-mirror/one-per-cluster.yaml workloads/rbd-mirror-stress-workunit.yaml}
- rbd/mirror/{base/install.yaml cluster/{2-node.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml rbd-mirror/one-per-cluster.yaml workloads/rbd-mirror-stress-workunit.yaml}
- SELinux denials found on ubuntu@smithi012.front.sepia.ceph.com: ['type=AVC msg=audit(1481253420.151:8286): avc: denied { open } for pid=11385 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481253937.158:8443): avc: denied { open } for pid=17743 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481253420.151:8286): avc: denied { read } for pid=11385 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481253688.221:8387): avc: denied { read } for pid=14751 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481253688.221:8387): avc: denied { open } for pid=14751 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481253937.158:8443): avc: denied { read } for pid=17743 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481253440.886:8326): avc: denied { read } for pid=12332 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481253440.886:8326): avc: denied { open } for pid=12332 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file']
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=10328195aad09f59d5c2c382bd9241c7418f744e TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin RBD_FEATURES=125 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd.sh' - }, 'changed': False, 'msg': 'Failed to update apt cache.'}}
- SELinux denials found on ubuntu@smithi003.front.sepia.ceph.com: ['type=AVC msg=audit(1481244833.591:4250): avc: denied { read } for pid=23876 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481244813.084:4167): avc: denied { read } for pid=23274 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481244833.591:4250): avc: denied { open } for pid=23876 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481244813.084:4167): avc: denied { open } for pid=23274 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file']
- SELinux denials found on ubuntu@smithi003.front.sepia.ceph.com: ['type=AVC msg=audit(1481245626.292:4358): avc: denied { open } for pid=18438 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481245483.209:4162): avc: denied { read } for pid=23241 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481245483.209:4162): avc: denied { open } for pid=23241 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481245626.292:4358): avc: denied { read } for pid=18438 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481245503.709:4257): avc: denied { open } for pid=23893 comm="ceph-osd" path="/var/lib/ceph/osd/ceph-0/type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file', 'type=AVC msg=audit(1481245503.709:4257): avc: denied { read } for pid=23893 comm="ceph-osd" name="type" dev="nvme0n1p2" ino=25 scontext=system_u:system_r:ceph_t:s0 tcontext=unconfined_u:object_r:unlabeled_t:s0 tclass=file']
- }, 'changed': False, 'msg': 'Failed to update apt cache.'}}
- 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3'
Re-running failed jobs:
#25 Updated by Loïc Dachary over 6 years ago
- Subject changed from jewel v10.2.5 to jewel v10.2.6
- Target version changed from v10.2.5 to v10.2.6
#26 Updated by Loïc Dachary over 6 years ago
- Description updated (diff)
#27 Updated by Loïc Dachary over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel..wip-jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 10661
- |\
- | + rgw: Have a flavor of bucket deletion to bypass GC and to trigger object deletions async.
- + Pull request 11470
- |\
- | + rgw:bucket check remove multipart prefix
- + Pull request 11476
- |\
- | + rgw: RGWCoroutinesManager::run returns status of last cr
- + Pull request 11477
- |\
- | + rgw: fix for assertion in RGWMetaSyncCR
- + Pull request 11497
- |\
- | + rgw: settle /info implementation across other swift-at-root features.
- | + swift /info implementation.
- | + rgw: add support for the healthcheck feature of Swift API.
- | + rgw: add support for the crossdomain.xml resource of Swift API.
- | + rgw: fix the handling of rgw_swift_url_prefix.
- | + rgw_http_errors: add http error code for 503
- | + rgw: Allow to serve Swift off the URL root
- + Pull request 11508
- |\
- | + utime.h: fix timezone issue in round_to_* funcs.
- + Pull request 11656
- |\
- | + ceph_volume_client: fix partial auth recovery
- | + ceph_volume_client: check if volume metadata is empty
- | + ceph_volume_client: fix _recover_auth_meta() method
- + Pull request 11759
- |\
- | + rgw: json encode/decode index_type, allow modification
- + Pull request 11866
- |\
- | + rgw: clean up RGWShardedOmapCRManager on early return
- + Pull request 11868
- |\
- | + rgw: store oldest mdlog period in rados
- + Pull request 11872
- |\
- | + rgw: delete entries_index in RGWFetchAllMetaCR
- + Pull request 11876
- |\
- | + rgw: fix the field 'total_time' of log entry in log show opt
- + Pull request 11947
- |\
- | + mon: update mon(peon)'s down_pending_out when osd up
- + Pull request 11990
- |\
- | + rgw: for the create_bucket api, if the input creation_time is zero, we should set it to 'now
- + Pull request 11991
- |\
- | + osd: limit omap data in push op
- + Pull request 11998
- |\
- | + FileStore::_do_fiemap: do not reference fiemap after it is freed
- | + test: add test for fiemap xfs issue when #extents > 1364
- | + FileStore:: fix fiemap issue in xfs when #extents > 1364
- + Pull request 12043
- |\
- | + rbd-mirror: fix sparse read optimization in image sync
- | + rbd-mirror: set SEQUENTIAL and NOCACHE advise flags on image sync
- | + rbd-mirror: Add sparse read for sync image
- + Pull request 12044
- |\
- | + rgw: fix for bucket delete racing with mdlog sync
- + Pull request 12045
- |\
- | + rgw_file: fix spurious mount entries w/Linux NFS client
- + Pull request 12079
- |\
- | + rgw: TempURL properly handles accounts created with the implicit tenant.
- + Pull request 12137
- |\
- | + client: fix stale entries in command table
- + Pull request 12147
- |\
- | + ceph-disk: enable --runtime ceph-osd systemd units
- | + build/ops: restart ceph-osd@.service after 20s instead of 100ms
- | + ceph-disk: trigger must ensure device ownership
- | + ceph-disk: systemd unit must run after local-fs.target
- + Pull request 12153
- |\
- | + mds: ignore 'session evict' when mds is replaying log
- + Pull request 12154
- |\
- | + mds: use projected path construction for access
- + Pull request 12155
- |\
- | + mds: require MAY_SET_POOL to set pool_ns
- + Pull request 12156
- |\
- | + rgw: look for region_map in rgw_region_root_pool
- | + rgw: region conversion respects pre-existing rgw_region_root_pool
- + Pull request 12210
- |\
- | + systemd/ceph-disk: reduce ceph-disk flock contention
- + Pull request 12239
- |\
- | + jewel: fixed the issue when --disable-server, compilation fails.
- + Pull request 12258
- |\
- | + rgw: add support for the prefix parameter in account listing of Swift API.
- | + rgw: optimize out ctor-copy in RGWListBuckets_ObjStore_SWIFT.
- + Pull request 12288
- |\
- | + doc, test: Add schemas for list-inconsistent-* rados command output
- | + test: Update testing for new list-inconsistent-obj output
- | + rados, osd: Improve attrs output of list-inconsistent-obj
- | + osd: Fix logging to help with diagnostics
- | + test: Fix use of wait_for_clean()
- | + common: Change cleanbin() to use base64 encoding, update ceph-objectstore-tool
- | + common: Move cleanbin() function to common/util.cc
- | + test: Add test support for deep-scrub
- | + common: Fix indentation
- | + osd: Handle corrupt attributes in get_object_context()
- | + test/osd-scrub-repair.sh: Use test case specific object names to help with diagnostics
- + Pull request 12291
- |\
- | + msg/simple/Pipe: handle addr decode error
- + Pull request 12302
- |\
- | + tests: check hostname --fqdn sanity before running cmake check
- | + tests: check hostname --fqdn sanity before running make check
- + Pull request 12313
- |\
- | + rgw: only set CURLOPT_UPLOAD for PUT/POST requests
- + Pull request 12314
- |\
- | + rgw: RGWBucketSyncStatusManager uses existing async_rados
- + Pull request 12315
- |\
- | + rgw: fix missing master zone for a single zone zonegroup
- + Pull request 12316
- |\
- | + rgw: add recovery procedure for upgrade to older version of jewel
- + Pull request 12320
- |\
- | + rgw: fix for versioned delete_multi_object
- | + rgw:fix for deleting objects name beginning and ending with underscores of one bucket using POST method of AWS's js sdk. Fixes: http://tracker.ceph.com/issues/17888
- + Pull request 12321
- |\
- | + qa/workunits/rbd: use image id when probing for image presence
- + Pull request 12322
- |\
- | + librbd: diffs to clone's first snapshot should include parent diffs
- + Pull request 12323
- |\
- | + librbd: account m_processing when failing request after refresh
- + Pull request 12324
- |\
- | + mds: force client flush snap data before truncating objects
- + Pull request 12325
- |\
- | + ceph_volume_client: set an existing auth ID's default mon caps
- + Pull request 12380
- |\
- | + os/ObjectStore: properly clone object map when replaying OP_COLL_MOVE_RENAME
- | + os/ObjectStore: properly clear object map when replaying OP_REMOVE
- + Pull request 12397
- |\
- | + rgw: do not abort when accept a CORS request with short origin
- + Pull request 12405
- |\
- | + install-deps.sh: unify indentation in case statement
- | + install-deps.sh: allow building on SLES systems
- | + build/ops: fix /etc/os-release parsing in install-deps.sh
- | + install-deps.sh: initial distro detection based on /etc/os-release
- + Pull request 12416
- |\
- | + msg: don't truncate message sequence to 32-bits
- + Pull request 12419
- |\
- | + rgw: omap_get_all() fixes
- | + rgw/rgw_rados: do not omap_getvals with (u64)-1 max
- + Pull request 12424
- |\
- | + qa/workunits/rbd: removed qemu-iotest case 077
- + Pull request 12425
- |\
- | + rbd-mirror: make 'rbd mirror image resync' work after split-brain
- | + qa/workunits/rbd: test_status_in_pool_dir: explicitly check grep return value
- | + rbd-mirror: split-brain issues should be clearly visible in mirror status
- | + rbd-mirror: fix gmock warnings in bootstrap request unit tests
- + Pull request 12426
- |\
- | + rbd: --max_part and --nbds_max options for nbd map
- + Pull request 12428
- |\
- | + radosgw-admin: 'zone placement modify' doesnt require pool names
- | + radosgw-admin: add 'zonegroup placement default' command
- | + radosgw-admin: fix 'placment' typos
- | + rgw_admin: commands to manage placement targets
- | + rgw-admin: add commands to manage zonegroup placement fields
- | + rgw: use set for zonegroup placement target tags
- + Pull request 12529
- |\
- | + rbd: utilize new API methods for image id and block name prefix
- | + librbd: new API methods to retrieve image id and block name prefix
- + Pull request 12542
- |\
- | + rgw: Replacing '+' with %20 in canonical uri for s3 v4 auth.
- + Pull request 12622
- |\
- | + rgw: log name instead of id for SystemMetaObj on failure
- | + rgw: drop unnecessary spacing in rgw zg init log
- + Pull request 12649
- |\
- | + librbd/diff_iterator: use proper snap to query parent overlap
- + Pull request 12677
- |\
- | + mon: OSDMonitor: trigger an immediate propose if any newly down osd is detected during tick()
- + Pull request 12678
- |\
- | + rgw: ldap: simple_bind() should set ldap option on tldap
- + Pull request 12686
- |\
- | + tests: rbd/test_lock_fence.sh: fix rbdrw.py relative path
- | + qa/tasks/workunit: clear clone dir before retrying checkout
- | + qa/tasks/workunit: retry on ceph.git if checkout fails
- | + qa/workunits: include extension for nose tests
- | + qa/workunits: use relative path instead of wget from git
- | + qa/tasks/workunit.py: add CEPH_BASE env var
- | + qa/tasks/workunit: leave workunits inside git checkout
- + Pull request 12738
- |\
- | + rgw: use explicit flag to cancel RGWCoroutinesManager::run()
- + Pull request 12739
- |\
- | + journal: prevent repetitive error messages after being blacklisted
- | + journal: avoid logging an error when a watch is blacklisted
- + Pull request 12741
- |\
- | + rbd: fix json formatting for image and journal status output
- + Pull request 12745
- |\
- | + tests: use ceph-jewel branch for s3tests
- + Pull request 12753
- |\
- | + librbd: ignore error when object map is already locked by current client
- + Pull request 12754
- |\
- | + rbd-nbd: support partition for rbd-nbd mapped raw block device.
- + Pull request 12755
- |\
- | + rados: optionally support reading omap key from file
- + Pull request 12756
- |\
- | + rbd-nbd: invalid error code for failed to read nbd request messages
- + Pull request 12761
- |\
- | + qa/tasks/admin_socket: subst in repo name
- + Pull request 12764
- |\
- | + rgw: fix decoding of creation_time and last_update.
- + Pull request 12783
- |\
- | + client: don't use special faked-up inode for /..
- + Pull request 12789
- |\
- | + osd: improve error message when FileStore op fails due to EPERM
- + Pull request 12822
- |\
- | + tests: subst repo and branch in qemu test urls
- | + tests: subst branch and repo in qa/tasks/qemu.py
- | + tests: subst repo name in krbd/unmap/tasks/unmap.yaml
- | + tests: subst repo name in qa/tasks/cram.py
- | + cram: support fetching from sha1 branch, tag, commit hash
- + Pull request 12836
- |\
- | + qa/tasks: add test_corrupt_backtrace
- | + mds: check for errors decoding backtraces
- + Pull request 12875
- |\
- | + osd/PG: publish PG stats when backfill-related states change
- + Pull request 12890
- + librados: blacklist_add should wait for latest OSD map
- + librbd: prevent assertion failure when journal IO is blacklisted
- + librbd: ignore blacklist error when releasing exclusive lock
- + librbd: fail immediately if the exclusive lock cannot be acquired
- + librbd: add new lock_get_owners / lock_break_lock API methods
- + librbd: separate break lock logic into standalone state machine
- + librbd: separate locker query into standalone state machine
- + librbd/exclusive_lock/AcquireRequest.cc: init lock_type
- + librbd: API methods to directly acquire and release the exclusive lock
- + Merge branch 'wip-17134-jewel' of https://github.com/dillaman/ceph into wip-18453-jewel
- + rbd-mirror: fix error messages formatting
- + librbd: ignore partial refresh error when acquiring exclusive lock
- + librbd: potential seg fault when blacklisting an image client
- + librbd: potential double-unwatch of watch handle upon error
- + librbd: deadlock when replaying journal during image open
- + librbd: improve image state machine debug log messages
- + librbd: remove unused refresh request logic
- + librbd: interlock image refresh and lock operations
- + librbd: image state machine now has hooks for lock requests
- + librbd: integrate asynchronous image rewatch state machine
- + librbd: helper state machine for asynchronous watch recovery
- + librbd: exclusive lock now supports reacquiring a lost lock
- + librbd: store exclusive lock cookie instead of recalculating
- + librbd: helper state machine to update lock cookie
- + cls_lock: support updating the lock cookie without releasing the lock
#28 Updated by Loïc Dachary over 6 years ago
rados¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 50)/50 --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-12_15:26:07-rados-wip-jewel-backports-distro-basic-smithi/
- **
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=hammer TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh' - "2017-01-14 09:13:29.649276 osd.3 172.21.15.44:6800/865390 12 : cluster [ERR] 5.0 shard 3: soid 5:a0216fbc:::repair_test_obj:head size 1 != size 223 from auth oi 5:a0216fbc:::repair_test_obj:head(20'1 client.4352.0:1 dirty|data_digest|omap_digest s 223 uv 1 dd 9a3a59aa od ffffffff)" in cluster log
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&ref=v0.80.8
- Command crashed: 'CEPH_CLIENT_ID=0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph_test_rados --max-ops 4000 --objects 500 --max-in-flight 16 --size 4000000 --min-stride-size 400000 --max-stride-size 800000 --max-seconds 0 --op snap_remove 50 --op snap_create 50 --op rollback 50 --op read 100 --op copy_from 50 --op write 50 --op write_excl 50 --op delete 50 --pool base'
- 'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=9d446bd416c52cd785ccf048ca67737ceafcdd7f
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rados/test_python.sh' - "wget
q -O /home/ubuntu/cephtest/admin_socket_client.0/objecter_requests -'http://git.ceph.com/?p=ceph.git;a=blob_plain;f=src/test/admin_socket/objecter_requests;hb=wip-jewel-backports' && chmod u=rx -- /home/ubuntu/cephtest/admin_socket_client.0/objecter_requests"- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/few.yaml msgr/simple.yaml rados.yaml thrashers/pggrow.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/few.yaml msgr/simple.yaml rados.yaml thrashers/default.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml rados.yaml thrashers/mapgap.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml rados.yaml thrashers/morepggrow.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml rados.yaml thrashers/pggrow.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/osd-delay.yaml msgr/simple.yaml rados.yaml thrashers/default.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/fastclose.yaml msgr/simple.yaml rados.yaml thrashers/mapgap.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/fastclose.yaml msgr/simple.yaml rados.yaml thrashers/morepggrow.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/fastclose.yaml msgr/simple.yaml rados.yaml thrashers/pggrow.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/fastclose.yaml msgr/simple.yaml rados.yaml thrashers/default.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/few.yaml msgr/async.yaml rados.yaml thrashers/mapgap.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/normal_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/few.yaml msgr/async.yaml rados.yaml thrashers/morepggrow.yaml workloads/admin_socket_objecter_requests.yaml}
- rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml hobj-sort.yaml msgr-failures/few.yaml msgr/async.yaml rados.yaml thrashers/pggrow.yaml workloads/admin_socket_objecter_requests.yaml}
- saw valgrind issues
Re-running failed jobs
#29 Updated by Loïc Dachary over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --suite-branch jewel --ceph wip-jewel-backports --machine-type vps --priority 1000
#30 Updated by Loïc Dachary over 6 years ago
Upgrade¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x --suite-branch jewel --ceph wip-jewel-backports --machine-type vps --priority 1000
Re-running dead job
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --suite-branch jewel --ceph wip-jewel-backports --machine-type vps --priority 1000
- fail http://pulpito.ceph.com/loic-2017-01-12_15:38:04-upgrade:hammer-x-wip-jewel-backports-distro-basic-vps
- "sudo yum -y install '' ceph-radosgw"
- 'git clone git://git.ceph.com/ceph-ci.git /home/ubuntu/cephtest/clone.client.0 ; cd -- /home/ubuntu/cephtest/clone.client.0 && git checkout firefly && mv qa/workunits /home/ubuntu/cephtest/workunit.client.0'
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=95cefea9fd9ab740263bf8bb4796fd864d9afe2b
- 'git clone git://git.ceph.com/ceph-ci.git /home/ubuntu/cephtest/clone.client.0 ; cd -- /home/ubuntu/cephtest/clone.client.0 && git checkout hammer && mv qa/workunits /home/ubuntu/cephtest/workunit.client.0'
- upgrade:hammer-x/parallel/{0-cluster/start.yaml 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 4-jewel.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_14.04.yaml}
- upgrade:hammer-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_14.04.yaml}
- upgrade:hammer-x/parallel/{0-cluster/start.yaml 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-osd-mds-mon.yaml 4-jewel.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/centos_7.2.yaml}
- upgrade:hammer-x/parallel/{0-cluster/start.yaml 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-osd-mds-mon.yaml 4-jewel.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/ubuntu_14.04.yaml}
- upgrade:hammer-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/centos_7.2.yaml}
- upgrade:hammer-x/parallel/{0-cluster/start.yaml 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-workload/{blogbench.yaml ec-rados-default.yaml rados_api.yaml rados_loadgenbig.yaml test_rbd_api.yaml test_rbd_python.yaml} 3-upgrade-sequence/upgrade-all.yaml 4-jewel.yaml 5-final-workload/{blogbench.yaml rados-snaps-few-objects.yaml rados_loadgenmix.yaml rados_mon_thrash.yaml rbd_cls.yaml rbd_import_export.yaml rgw_swift.yaml} distros/centos_7.2.yaml}
Re-running failed tests
#31 Updated by Loïc Dachary over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports --suite-branch jewel -k distro -m smithi -s powercycle -p 1000 --email loic@dachary.org
- fail http://pulpito.ceph.com/loic-2017-01-12_15:39:38-powercycle-wip-jewel-backports-distro-basic-smithi
- "wget
q -O /home/ubuntu/cephtest/admin_socket_client.0/objecter_requests -'http://git.ceph.com/?p=ceph.git;a=blob_plain;f=src/test/admin_socket/objecter_requests;hb=wip-jewel-backports' && chmod u=rx -- /home/ubuntu/cephtest/admin_socket_client.0/objecter_requests" - timed out waiting for admin_socket to appear after osd.0 restart
- "wget
Re-running failed jobs
#32 Updated by Loïc Dachary over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-12_15:40:36-fs-wip-jewel-backports-distro-basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/libcephfs-java/test.sh' - Test failure: test_session_reject (tasks.cephfs.test_sessionmap.TestSessionMap)
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/fs/test_python.sh'- fs/basic/{clusters/fixed-2-ucephfs.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml fs/btrfs.yaml inline/no.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_python.yaml}
- fs/basic/{clusters/fixed-2-ucephfs.yaml debug/mds_client.yaml dirfrag/frag_enable.yaml fs/btrfs.yaml inline/yes.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_python.yaml}
- saw valgrind issues
- fs/verify/{clusters/fixed-2-ucephfs.yaml debug/{mds_client.yaml mon.yaml} dirfrag/frag_enable.yaml fs/btrfs.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/libcephfs_interface_tests.yaml validater/valgrind.yaml}
- fs/verify/{clusters/fixed-2-ucephfs.yaml debug/{mds_client.yaml mon.yaml} dirfrag/frag_enable.yaml fs/btrfs.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/cfuse_workunit_suites_dbench.yaml validater/valgrind.yaml}
- fs/verify/{clusters/fixed-2-ucephfs.yaml debug/{mds_client.yaml mon.yaml} dirfrag/frag_enable.yaml fs/btrfs.yaml overrides/whitelist_wrongly_marked_down.yaml tasks/cfuse_workunit_suites_fsstress.yaml validater/valgrind.yaml}
- 'mkdir
Re-running failed jobs
#33 Updated by Loïc Dachary over 6 years ago
rgw¶
teuthology-suite -k distro --priority 1000 --suite rgw --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-12_15:41:47-rgw-wip-jewel-backports-distro-basic-smithi
- saw valgrind issues
- rgw/verify/{clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/replicated.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec-profile.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/replicated.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/replicated.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec-profile.yaml tasks/rgw_s3tests_multiregion.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/apache.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec-cache.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/replicated.yaml tasks/rgw_swift.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/ec-profile.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- rgw/verify/{clusters/fixed-2.yaml frontend/civetweb.yaml fs/btrfs.yaml msgr-failures/few.yaml overrides.yaml rgw_pool_type/replicated.yaml tasks/rgw_s3tests.yaml validater/valgrind.yaml}
- saw valgrind issues
Re-running failed jobs
- fail http://pulpito.ceph.com/loic-2017-01-23_06:30:03-rgw-wip-jewel-backports-distro-basic-smithi
- three passed, all the rest failed due to "valgrind issues"
#34 Updated by Loïc Dachary over 6 years ago
rbd¶
teuthology-suite --priority 1000 --suite rbd --suite-branch jewel --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-12_15:43:54-rbd-wip-jewel-backports---basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 6h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror_stress.sh' - 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin RBD_FEATURES=61 VALGRIND=memcheck adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh' - 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin RBD_FEATURES=125 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh'- rbd/librbd/{cache/writeback.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/none.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/none.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/none.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- rbd/librbd/{cache/none.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_journaling.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin RBD_FEATURES=61 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh'- rbd/librbd/{cache/none.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/none.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/none.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/none.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests_with_defaults.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh'- rbd/basic/{base/install.yaml cachepool/none.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/many.yaml tasks/rbd_python_api_tests_old_format.yaml}
- rbd/basic/{base/install.yaml cachepool/small.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml tasks/rbd_python_api_tests_old_format.yaml}
- rbd/basic/{base/install.yaml cachepool/small.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/many.yaml tasks/rbd_python_api_tests_old_format.yaml}
- rbd/basic/{base/install.yaml cachepool/none.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml tasks/rbd_python_api_tests_old_format.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin RBD_FEATURES=1 VALGRIND=memcheck adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh' - 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin RBD_FEATURES=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh'- rbd/librbd/{cache/writethrough.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/none.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/none.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/small.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/none.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/on.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/writethrough.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/none.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- rbd/librbd/{cache/writeback.yaml cachepool/none.yaml clusters/{fixed-3.yaml openstack.yaml} copy-on-read/off.yaml fs/xfs.yaml msgr-failures/few.yaml workloads/python_api_tests.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_lock_fence.sh'- rbd/basic/{base/install.yaml cachepool/small.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml tasks/rbd_lock_and_fence.yaml}
- rbd/basic/{base/install.yaml cachepool/none.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/many.yaml tasks/rbd_lock_and_fence.yaml}
- rbd/basic/{base/install.yaml cachepool/none.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/few.yaml tasks/rbd_lock_and_fence.yaml}
- rbd/basic/{base/install.yaml cachepool/small.yaml clusters/{fixed-1.yaml openstack.yaml} fs/xfs.yaml msgr-failures/many.yaml tasks/rbd_lock_and_fence.yaml}
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/qemu-iotests.sh' - 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.cluster1.client.mirror/rbd/rbd_mirror.sh' - 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=87836ae7a2a75a9c7324de4bf91a2e8c27058832 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin RBD_FEATURES=125 VALGRIND=memcheck adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/workunit.client.0/rbd/test_librbd_python.sh'
- 'mkdir
Re-running failed jobs
#35 Updated by Loïc Dachary over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel..wip-jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 10661
- |\
- | + rgw: Have a flavor of bucket deletion to bypass GC and to trigger object deletions async.
- + Pull request 11470
- |\
- | + rgw:bucket check remove multipart prefix
- + Pull request 11476
- |\
- | + rgw: RGWCoroutinesManager::run returns status of last cr
- + Pull request 11477
- |\
- | + rgw: fix for assertion in RGWMetaSyncCR
- + Pull request 11497
- |\
- | + rgw: settle /info implementation across other swift-at-root features.
- | + swift /info implementation.
- | + rgw: add support for the healthcheck feature of Swift API.
- | + rgw: add support for the crossdomain.xml resource of Swift API.
- | + rgw: fix the handling of rgw_swift_url_prefix.
- | + rgw_http_errors: add http error code for 503
- | + rgw: Allow to serve Swift off the URL root
- + Pull request 11508
- |\
- | + utime.h: fix timezone issue in round_to_* funcs.
- + Pull request 11759
- |\
- | + rgw: json encode/decode index_type, allow modification
- + Pull request 11866
- |\
- | + rgw: clean up RGWShardedOmapCRManager on early return
- + Pull request 11868
- |\
- | + rgw: store oldest mdlog period in rados
- + Pull request 11872
- |\
- | + rgw: delete entries_index in RGWFetchAllMetaCR
- + Pull request 11876
- |\
- | + rgw: fix the field 'total_time' of log entry in log show opt
- + Pull request 11947
- |\
- | + mon: update mon(peon)'s down_pending_out when osd up
- + Pull request 11990
- |\
- | + rgw: for the create_bucket api, if the input creation_time is zero, we should set it to 'now
- + Pull request 11991
- |\
- | + osd: limit omap data in push op
- + Pull request 11998
- |\
- | + FileStore::_do_fiemap: do not reference fiemap after it is freed
- | + test: add test for fiemap xfs issue when #extents > 1364
- | + FileStore:: fix fiemap issue in xfs when #extents > 1364
- + Pull request 12044
- |\
- | + rgw: fix for bucket delete racing with mdlog sync
- + Pull request 12045
- |\
- | + rgw_file: fix spurious mount entries w/Linux NFS client
- + Pull request 12079
- |\
- | + rgw: TempURL properly handles accounts created with the implicit tenant.
- + Pull request 12153
- |\
- | + mds: ignore 'session evict' when mds is replaying log
- + Pull request 12156
- |\
- | + rgw: look for region_map in rgw_region_root_pool
- | + rgw: region conversion respects pre-existing rgw_region_root_pool
- + Pull request 12239
- |\
- | + jewel: fixed the issue when --disable-server, compilation fails.
- + Pull request 12258
- |\
- | + rgw: add support for the prefix parameter in account listing of Swift API.
- | + rgw: optimize out ctor-copy in RGWListBuckets_ObjStore_SWIFT.
- + Pull request 12291
- |\
- | + msg/simple/Pipe: handle addr decode error
- + Pull request 12302
- |\
- | + tests: check hostname --fqdn sanity before running cmake check
- | + tests: check hostname --fqdn sanity before running make check
- + Pull request 12313
- |\
- | + rgw: only set CURLOPT_UPLOAD for PUT/POST requests
- + Pull request 12314
- |\
- | + rgw: RGWBucketSyncStatusManager uses existing async_rados
- + Pull request 12315
- |\
- | + rgw: fix missing master zone for a single zone zonegroup
- + Pull request 12316
- |\
- | + rgw: add recovery procedure for upgrade to older version of jewel
- + Pull request 12320
- |\
- | + rgw: fix for versioned delete_multi_object
- | + rgw:fix for deleting objects name beginning and ending with underscores of one bucket using POST method of AWS's js sdk. Fixes: http://tracker.ceph.com/issues/17888
- + Pull request 12321
- |\
- | + qa/workunits/rbd: use image id when probing for image presence
- + Pull request 12322
- |\
- | + librbd: diffs to clone's first snapshot should include parent diffs
- + Pull request 12323
- |\
- | + librbd: account m_processing when failing request after refresh
- + Pull request 12324
- |\
- | + mds: force client flush snap data before truncating objects
- + Pull request 12380
- |\
- | + os/ObjectStore: properly clone object map when replaying OP_COLL_MOVE_RENAME
- | + os/ObjectStore: properly clear object map when replaying OP_REMOVE
- + Pull request 12397
- |\
- | + rgw: do not abort when accept a CORS request with short origin
- + Pull request 12405
- |\
- | + install-deps.sh: unify indentation in case statement
- | + install-deps.sh: allow building on SLES systems
- | + build/ops: fix /etc/os-release parsing in install-deps.sh
- | + install-deps.sh: initial distro detection based on /etc/os-release
- + Pull request 12416
- |\
- | + msg: don't truncate message sequence to 32-bits
- + Pull request 12419
- |\
- | + rgw: omap_get_all() fixes
- | + rgw/rgw_rados: do not omap_getvals with (u64)-1 max
- + Pull request 12425
- |\
- | + rbd-mirror: make 'rbd mirror image resync' work after split-brain
- | + qa/workunits/rbd: test_status_in_pool_dir: explicitly check grep return value
- | + rbd-mirror: split-brain issues should be clearly visible in mirror status
- | + rbd-mirror: fix gmock warnings in bootstrap request unit tests
- + Pull request 12426
- |\
- | + rbd: --max_part and --nbds_max options for nbd map
- + Pull request 12428
- |\
- | + radosgw-admin: 'zone placement modify' doesnt require pool names
- | + radosgw-admin: add 'zonegroup placement default' command
- | + radosgw-admin: fix 'placment' typos
- | + rgw_admin: commands to manage placement targets
- | + rgw-admin: add commands to manage zonegroup placement fields
- | + rgw: use set for zonegroup placement target tags
- + Pull request 12529
- |\
- | + rbd: utilize new API methods for image id and block name prefix
- | + librbd: new API methods to retrieve image id and block name prefix
- + Pull request 12542
- |\
- | + rgw: Replacing '+' with %20 in canonical uri for s3 v4 auth.
- + Pull request 12622
- |\
- | + rgw: log name instead of id for SystemMetaObj on failure
- | + rgw: drop unnecessary spacing in rgw zg init log
- + Pull request 12649
- |\
- | + librbd/diff_iterator: use proper snap to query parent overlap
- + Pull request 12677
- |\
- | + mon: OSDMonitor: trigger an immediate propose if any newly down osd is detected during tick()
- + Pull request 12678
- |\
- | + rgw: ldap: simple_bind() should set ldap option on tldap
- + Pull request 12738
- |\
- | + rgw: use explicit flag to cancel RGWCoroutinesManager::run()
- + Pull request 12739
- |\
- | + journal: prevent repetitive error messages after being blacklisted
- | + journal: avoid logging an error when a watch is blacklisted
- + Pull request 12741
- |\
- | + rbd: fix json formatting for image and journal status output
- + Pull request 12753
- |\
- | + librbd: ignore error when object map is already locked by current client
- + Pull request 12754
- |\
- | + rbd-nbd: support partition for rbd-nbd mapped raw block device.
- + Pull request 12755
- |\
- | + rados: optionally support reading omap key from file
- + Pull request 12756
- |\
- | + rbd-nbd: invalid error code for failed to read nbd request messages
- + Pull request 12761
- |\
- | + qa/tasks/admin_socket: subst in repo name
- + Pull request 12764
- |\
- | + rgw: fix decoding of creation_time and last_update.
- + Pull request 12783
- |\
- | + client: don't use special faked-up inode for /..
- + Pull request 12789
- |\
- | + osd: improve error message when FileStore op fails due to EPERM
- + Pull request 12822
- |\
- | + tests: subst repo and branch in qemu test urls
- | + tests: subst branch and repo in qa/tasks/qemu.py
- | + tests: subst repo name in krbd/unmap/tasks/unmap.yaml
- | + tests: subst repo name in qa/tasks/cram.py
- | + cram: support fetching from sha1 branch, tag, commit hash
- + Pull request 12875
- |\
- | + osd/PG: publish PG stats when backfill-related states change
- + Pull request 12890
- |\
- | + librados: blacklist_add should wait for latest OSD map
- | + librbd: prevent assertion failure when journal IO is blacklisted
- | + librbd: ignore blacklist error when releasing exclusive lock
- | + librbd: fail immediately if the exclusive lock cannot be acquired
- | + librbd: add new lock_get_owners / lock_break_lock API methods
- | + librbd: separate break lock logic into standalone state machine
- | + librbd: separate locker query into standalone state machine
- | + librbd/exclusive_lock/AcquireRequest.cc: init lock_type
- | + librbd: API methods to directly acquire and release the exclusive lock
- | + Merge branch 'wip-17134-jewel' of https://github.com/dillaman/ceph into wip-18453-jewel
- | + rbd-mirror: fix error messages formatting
- | + librbd: ignore partial refresh error when acquiring exclusive lock
- | + librbd: potential seg fault when blacklisting an image client
- | + librbd: potential double-unwatch of watch handle upon error
- | + librbd: deadlock when replaying journal during image open
- | + librbd: improve image state machine debug log messages
- | + librbd: remove unused refresh request logic
- | + librbd: interlock image refresh and lock operations
- | + librbd: image state machine now has hooks for lock requests
- | + librbd: integrate asynchronous image rewatch state machine
- | + librbd: helper state machine for asynchronous watch recovery
- | + librbd: exclusive lock now supports reacquiring a lost lock
- | + librbd: store exclusive lock cookie instead of recalculating
- | + librbd: helper state machine to update lock cookie
- | + cls_lock: support updating the lock cookie without releasing the lock
- + Pull request 12909
- |\
- | + librbd: block concurrent in-flight object map updates for the same object
- | + librbd: new block guard helper to prevent concurrent IO to blocks
- | + librbd: convert ObjectMap to template for unit testing
- | + librbd: clean up object map update interface
- | + librbd: update in-memory object map after on-disk update committed
- + Pull request 12917
- |\
- | + rgw: Handle multiple listening addreses w/ optional ssl correctly with civetweb.
- | + rgw: s3: secure_port should override port, also apply ssl default right.
- | + rgw: Get civetweb ssl enhancement: wip-listen3 = mg_get_local_addr
- | + rgw: Document that radosgw now supports SSL.
- | + rgw: civetweb/openssl: automagic: load libssl.so and libcrypto.so by soname.
- | + rgw: civetweb/openssl: Load libssl.so and libcrypto.so by soname.
- | + rgw: cmake: remove useless civetweb include path side effect.
- + Pull request 12921
- |\
- | + mds/server: skip unwanted dn in handle_client_readdir
- + Pull request 12978
- |\
- | + ReplicatedBackend: take read locks for clone sources during recovery
- + Pull request 12997
- |\
- | + rgw_rados: add guard assert in add_io()
- | + rgw_rados: sanitize dout print in GWRados::get_obj_iterate_cb
- + Pull request 13001
- |\
- | + rgw: RGWAsyncRadosRequest drops notifier ref on cancel
- | + rgw: remove circular reference in RGWAsyncRadosRequest
- | + rgw: release RGWAioCompletionNotifier refs on destruction
- + Pull request 13025
- |\
- | + Ceph-disk to use correct user in check_journal_req
- + Pull request 13029
- |\
- | + client/Client.cc: prevent segfaulting
- + Pull request 13040
- |\
- | + tests: run fs/thrash on xfs instead of btrfs
- + Pull request 13043
- |\
- | + Doc: Fixes Python Swift client commands
- + Pull request 13045
- |\
- | + mon: do not send duplicated osdmap msg to not sync'ed osd
- + Pull request 13047
- |\
- | + test: Add test support for deep-scrub
- | + osd: Add trigger_scrub admin socket command
- | + test: Add test for keeping deep-scrub information
- | + osd: When deep-scrub errors present upgrade regular scrubs
- + Pull request 13048
- |\
- | + selinux: Allow ceph to manage tmp files
- + Pull request 13049
- |\
- | + doc: document hostname constraints for rados bench
- + Pull request 13050
- |\
- | + qa/distros: centos_7.yaml -> centos.yaml
- | + qa/suites: centos_7.2.yaml -> centos_7.yaml
- | + qa/distros: add centos 7.3
- | + qa/distros: add centos 7 yaml; use that instead
- + Pull request 13062
- |\
- | + msg/simple: clear_pipe when wait() is mopping up pipes
- + Pull request 13104
- + tasks/rbd_fio: unmap rbd devices on cleanup
- + tasks/rbd_fio: don't use sudo unnecessarily
#36 Updated by Loïc Dachary over 6 years ago
Upgrade¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x --ceph wip-jewel-backports --machine-type vps --priority 1000
- cannot be scheduled
- IOError: /home/smithfarm/src/git.ceph.com_ceph-c_wip-jewel-backports/qa/suites/upgrade/jewel-x/point-to-point-x/distros/centos.yaml does not exist (abs /home/smithfarm/src/git.ceph.com_ceph-c_wip-jewel-backports/qa/suites/upgrade/jewel-x/point-to-point-x/distros/centos.yaml)
- added a commit to https://github.com/ceph/ceph/pull/13050
- pushed wip-18406-jewel to Shaman so we can see if this one commit is sufficient or if more are needed
- IOError: /home/smithfarm/src/git.ceph.com_ceph-c_wip-jewel-backports/qa/suites/upgrade/jewel-x/point-to-point-x/distros/centos.yaml does not exist (abs /home/smithfarm/src/git.ceph.com_ceph-c_wip-jewel-backports/qa/suites/upgrade/jewel-x/point-to-point-x/distros/centos.yaml)
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --ceph wip-jewel-backports --machine-type vps --priority 1000
- fail http://pulpito.ceph.com/loic-2017-01-26_22:13:59-upgrade:hammer-x-wip-jewel-backports-distro-basic-vps
- known bug http://tracker.ceph.com/issues/18089
- all the upgrade:hammer-x/f-h-x-offline jobs that got run on CentOS
- opened https://github.com/ceph/ceph/pull/13153 to fix
- known bug http://tracker.ceph.com/issues/18089
#37 Updated by Loïc Dachary over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-jewel-backports --machine-type vps --priority 1000
- fail http://pulpito.ceph.com/loic-2017-01-26_22:12:12-ceph-disk-wip-jewel-backports-distro-basic-vps
- infrastructure noise Downburst failed on ubuntu@vpm051.front.sepia.ceph.com: libvirt: XML-RPC error : Cannot recv data: ssh: connect to host
Re-running the failed job:
- fail http://pulpito.ceph.com/smithfarm-2017-01-27_14:51:25-ceph-disk-wip-jewel-backports-distro-basic-vps/
- known bug http://tracker.ceph.com/issues/18416
- fixed by cherry-picking a patch into https://github.com/ceph/ceph/pull/13050
- known bug http://tracker.ceph.com/issues/18416
#38 Updated by Loïc Dachary over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports -k distro -m smithi -s powercycle -p 1000 --email loic@dachary.org
- fail http://pulpito.ceph.com/loic-2017-01-26_22:11:16-powercycle-wip-jewel-backports-distro-basic-smithi
- three jobs failed, out of 28
Re-running 3 failed jobs:
#39 Updated by Loïc Dachary over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-26_22:04:20-fs-wip-jewel-backports-distro-basic-smithi
- saw valgrind issues (3 failed jobs)
- cephfs_java test failure (1 failed job; can probably be ignored)
Re-running 4 failed jobs:
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-28_11:44:02-fs-wip-jewel-backports-distro-basic-smithi/
- saw valgrind issues (2 failed jobs)
#40 Updated by Loïc Dachary over 6 years ago
rgw¶
teuthology-suite -k distro --priority 1000 --suite rgw --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-26_22:03:35-rgw-wip-jewel-backports-distro-basic-smithi
- "saw valgrind issues" new bug http://tracker.ceph.com/issues/18744 (the daemons in the notcmalloc build are linked with libtcmalloc) (17 failed jobs)
- clock skew (1 failed job)
Re-running 18 failed jobs:
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-28_11:40:26-rgw-wip-jewel-backports-distro-basic-smithi/ - 11 failed jobs
- "saw valgrind issues" new bug http://tracker.ceph.com/issues/18744 (the daemons in the notcmalloc build are linked with libtcmalloc) (11 failed jobs)
Found and staged pending RGW valgrind-related backports:
- https://github.com/ceph/ceph/pull/13163
- https://github.com/ceph/ceph/pull/13178
- https://github.com/ceph/ceph/pull/13179
Hopefully that will clear up the valgrind issues in the next integration run.
NEWS FLASH rgw valgrind issues are appearing also in hammer QE testing, but only on centos. Hammer testing is done on CentOS 7.3 and Ubuntu 14.04 only. Upon closer examination, I don't see Ubuntu 14.04 in any of the jobs with valgrind issues.
#41 Updated by Loïc Dachary over 6 years ago
rbd¶
teuthology-suite --priority 1000 --suite rbd --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-26_22:02:45-rbd-wip-jewel-backports---basic-smithi
- known bug, possible regression?? http://tracker.ceph.com/issues/17695 supposedly already fixed in jewel long ago by https://github.com/ceph/ceph/pull/11644
Re-running the one failed job:
#42 Updated by Loïc Dachary over 6 years ago
rados¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 50)/50 --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-01-26_22:01:29-rados-wip-jewel-backports-distro-basic-smithi
- known bug http://tracker.ceph.com/issues/18089 - opened https://github.com/ceph/ceph/pull/13170 to drop these two problematic tests
- saw valgrind issues
- rados/verify/{1thrash/default.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/async.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml}
- rados/verify/{1thrash/none.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml rados.yaml tasks/rados_api_tests.yaml validater/valgrind.yaml}
- Regular scrub request, losing deep-scrub details" in cluster log - already fixed by adding a commit to https://github.com/ceph/ceph/pull/13047
- new bug http://tracker.ceph.com/issues/18719 - cluster gets stuck in "HEALTH_WARN all OSDs are running jewel or later but the 'require_jewel_osds' osdmap flag is not set" after upgrading from hammer
- infrastructure noise Could not reconnect to ubuntu@smithi016.front.sepia.ceph.com (host failed to come back from 'sudo shutdown -r now')
- clock skew mon.0 172.21.15.38:6789/0 5 : cluster [WRN] message from mon.2 was stamped 9.536220s in the future, clocks not synchronized" in cluster log
- first dead job restart of last upgraded MON never completes
- second dead job rados task (under thrashosds) never completes, log output suddenly stops and job is killed 10 hours later
- third dead job rados task (under thrashosds) fails to complete before thrashosds timeout expires
Re-running the jobs that stand a chance of succeeding on re-run:
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-28_11:36:51-rados-wip-jewel-backports-distro-basic-smithi/
- saw valgrind issues (1 failed job)
- problem with hammer upgrade rados/upgrade/{hammer-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{ec-rados-plugin=jerasure-k=3-m=1.yaml rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml test_cache-pool-snaps.yaml}} rados.yaml}
Re-running with the fix from PR#13161
teuthology-suite -k distro --priority 101 --suite rados --email ncutler@suse.com --machine-type smithi --ceph wip-lfn-upgrade-hammer --filter="rados/upgrade/{hammer-x-singleton/{0-cluster/{openstack.yaml start.yaml} 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{ec-rados-plugin=jerasure-k=3-m=1.yaml rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml test_cache-pool-snaps.yaml}} rados.yaml}"
Re-running the upgrade jobs to test David Galloway's fix from http://tracker.ceph.com/issues/18089#note-12
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-29_02:56:52-rados-wip-jewel-backports-distro-basic-smithi/
- opened https://github.com/ceph/ceph/pull/13170 to drop these two problematic tests
#43 Updated by Nathan Cutler over 6 years ago
rados baseline (jewel)¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 50)/50 --email loic@dachary.org --ceph jewel --machine-type smithi --ceph-repo http://github.com/ceph/ceph.git --suite-repo http://github.com/ceph/ceph.git
- running http://pulpito.ceph.com:80/smithfarm-2017-01-29_03:27:21-rados-jewel-distro-basic-smithi/
- known bug http://tracker.ceph.com/issues/18089
- David Galloway is helping get the right builds on the chacra hosts
- re-testing the failed jobs at http://tracker.ceph.com/issues/18089
- valgrind issues
- selinux issue
- segfault
- known bug http://tracker.ceph.com/issues/18089
#44 Updated by Nathan Cutler over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel..wip-jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 10661
- |\
- | + rgw: Have a flavor of bucket deletion to bypass GC and to trigger object deletions async.
- + Pull request 11470
- |\
- | + rgw:bucket check remove multipart prefix
- + Pull request 11476
- |\
- | + rgw: RGWCoroutinesManager::run returns status of last cr
- + Pull request 11477
- |\
- | + rgw: fix for assertion in RGWMetaSyncCR
- + Pull request 11497
- |\
- | + rgw: settle /info implementation across other swift-at-root features.
- | + swift /info implementation.
- | + rgw: add support for the healthcheck feature of Swift API.
- | + rgw: add support for the crossdomain.xml resource of Swift API.
- | + rgw: fix the handling of rgw_swift_url_prefix.
- | + rgw_http_errors: add http error code for 503
- | + rgw: Allow to serve Swift off the URL root
- + Pull request 11508
- |\
- | + utime.h: fix timezone issue in round_to_* funcs.
- + Pull request 11759
- |\
- | + rgw: json encode/decode index_type, allow modification
- + Pull request 11866
- |\
- | + rgw: clean up RGWShardedOmapCRManager on early return
- + Pull request 11868
- |\
- | + rgw: store oldest mdlog period in rados
- + Pull request 11872
- |\
- | + rgw: delete entries_index in RGWFetchAllMetaCR
- + Pull request 11876
- |\
- | + rgw: fix the field 'total_time' of log entry in log show opt
- + Pull request 11947
- |\
- | + mon: update mon(peon)'s down_pending_out when osd up
- + Pull request 11990
- |\
- | + rgw: for the create_bucket api, if the input creation_time is zero, we should set it to 'now
- + Pull request 11991
- |\
- | + osd: limit omap data in push op
- + Pull request 11998
- |\
- | + FileStore::_do_fiemap: do not reference fiemap after it is freed
- | + test: add test for fiemap xfs issue when #extents > 1364
- | + FileStore:: fix fiemap issue in xfs when #extents > 1364
- + Pull request 12044
- |\
- | + rgw: fix for bucket delete racing with mdlog sync
- + Pull request 12045
- |\
- | + rgw_file: fix spurious mount entries w/Linux NFS client
- + Pull request 12079
- |\
- | + rgw: TempURL properly handles accounts created with the implicit tenant.
- + Pull request 12153
- |\
- | + mds: ignore 'session evict' when mds is replaying log
- + Pull request 12156
- |\
- | + rgw: look for region_map in rgw_region_root_pool
- | + rgw: region conversion respects pre-existing rgw_region_root_pool
- + Pull request 12239
- |\
- | + jewel: fixed the issue when --disable-server, compilation fails.
- + Pull request 12258
- |\
- | + rgw: add support for the prefix parameter in account listing of Swift API.
- | + rgw: optimize out ctor-copy in RGWListBuckets_ObjStore_SWIFT.
- + Pull request 12291
- |\
- | + msg/simple/Pipe: handle addr decode error
- + Pull request 12302
- |\
- | + tests: check hostname --fqdn sanity before running cmake check
- | + tests: check hostname --fqdn sanity before running make check
- + Pull request 12313
- |\
- | + rgw: only set CURLOPT_UPLOAD for PUT/POST requests
- + Pull request 12314
- |\
- | + rgw: RGWBucketSyncStatusManager uses existing async_rados
- + Pull request 12315
- |\
- | + rgw: fix missing master zone for a single zone zonegroup
- + Pull request 12316
- |\
- | + rgw: add recovery procedure for upgrade to older version of jewel
- + Pull request 12320
- |\
- | + rgw: fix for versioned delete_multi_object
- | + rgw:fix for deleting objects name beginning and ending with underscores of one bucket using POST method of AWS's js sdk. Fixes: http://tracker.ceph.com/issues/17888
- + Pull request 12322
- |\
- | + librbd: diffs to clone's first snapshot should include parent diffs
- + Pull request 12323
- |\
- | + librbd: account m_processing when failing request after refresh
- + Pull request 12324
- |\
- | + mds: force client flush snap data before truncating objects
- + Pull request 12380
- |\
- | + os/ObjectStore: properly clone object map when replaying OP_COLL_MOVE_RENAME
- | + os/ObjectStore: properly clear object map when replaying OP_REMOVE
- + Pull request 12397
- |\
- | + rgw: do not abort when accept a CORS request with short origin
- + Pull request 12405
- |\
- | + install-deps.sh: unify indentation in case statement
- | + install-deps.sh: allow building on SLES systems
- | + build/ops: fix /etc/os-release parsing in install-deps.sh
- | + install-deps.sh: initial distro detection based on /etc/os-release
- + Pull request 12416
- |\
- | + msg: don't truncate message sequence to 32-bits
- + Pull request 12419
- |\
- | + rgw: omap_get_all() fixes
- | + rgw/rgw_rados: do not omap_getvals with (u64)-1 max
- + Pull request 12425
- |\
- | + rbd-mirror: fix gmock warnings in bootstrap request unit tests
- | + qa/workunits/rbd: test_status_in_pool_dir: explicitly check grep return value
- | + rbd-mirror: make 'rbd mirror image resync' work after split-brain
- | + rbd-mirror: split-brain issues should be clearly visible in mirror status
- | + qa/workunits/rbd: use image id when probing for image presence
- | + qa/workunits/rbd: check status also in pool dir after asok commands
- | + qa/workunits/rbd: wait for image deleted before checking health
- | + qa/workunits/rbd: wait for image deleted before checking health
- | + qa/workunits/rbd: small fixup and improvements for rbd-mirror tests
- + Pull request 12426
- |\
- | + rbd: --max_part and --nbds_max options for nbd map
- + Pull request 12428
- |\
- | + radosgw-admin: 'zone placement modify' doesnt require pool names
- | + radosgw-admin: add 'zonegroup placement default' command
- | + radosgw-admin: fix 'placment' typos
- | + rgw_admin: commands to manage placement targets
- | + rgw-admin: add commands to manage zonegroup placement fields
- | + rgw: use set for zonegroup placement target tags
- + Pull request 12529
- |\
- | + rbd: utilize new API methods for image id and block name prefix
- | + librbd: new API methods to retrieve image id and block name prefix
- + Pull request 12542
- |\
- | + rgw: Replacing '+' with %20 in canonical uri for s3 v4 auth.
- + Pull request 12622
- |\
- | + rgw: log name instead of id for SystemMetaObj on failure
- | + rgw: drop unnecessary spacing in rgw zg init log
- + Pull request 12649
- |\
- | + librbd/diff_iterator: use proper snap to query parent overlap
- + Pull request 12677
- |\
- | + mon: OSDMonitor: trigger an immediate propose if any newly down osd is detected during tick()
- + Pull request 12678
- |\
- | + rgw: ldap: simple_bind() should set ldap option on tldap
- + Pull request 12738
- |\
- | + rgw: use explicit flag to cancel RGWCoroutinesManager::run()
- + Pull request 12739
- |\
- | + journal: prevent repetitive error messages after being blacklisted
- | + journal: avoid logging an error when a watch is blacklisted
- + Pull request 12741
- |\
- | + rbd: fix json formatting for image and journal status output
- + Pull request 12753
- |\
- | + librbd: ignore error when object map is already locked by current client
- + Pull request 12755
- |\
- | + rados: optionally support reading omap key from file
- + Pull request 12756
- |\
- | + rbd-nbd: invalid error code for failed to read nbd request messages
- + Pull request 12761
- |\
- | + qa/tasks/admin_socket: subst in repo name
- + Pull request 12764
- |\
- | + rgw: fix decoding of creation_time and last_update.
- + Pull request 12783
- |\
- | + client: don't use special faked-up inode for /..
- + Pull request 12789
- |\
- | + osd: improve error message when FileStore op fails due to EPERM
- + Pull request 12822
- |\
- | + tests: subst repo and branch in qemu test urls
- | + tests: subst branch and repo in qa/tasks/qemu.py
- | + tests: subst repo name in krbd/unmap/tasks/unmap.yaml
- | + tests: subst repo name in qa/tasks/cram.py
- | + cram: support fetching from sha1 branch, tag, commit hash
- + Pull request 12875
- |\
- | + osd/PG: publish PG stats when backfill-related states change
- + Pull request 12890
- |\
- | + librados: blacklist_add should wait for latest OSD map
- | + librbd: prevent assertion failure when journal IO is blacklisted
- | + librbd: ignore blacklist error when releasing exclusive lock
- | + librbd: fail immediately if the exclusive lock cannot be acquired
- | + librbd: add new lock_get_owners / lock_break_lock API methods
- | + librbd: separate break lock logic into standalone state machine
- | + librbd: separate locker query into standalone state machine
- | + librbd/exclusive_lock/AcquireRequest.cc: init lock_type
- | + librbd: API methods to directly acquire and release the exclusive lock
- | + Merge branch 'wip-17134-jewel' of https://github.com/dillaman/ceph into wip-18453-jewel
- | + rbd-mirror: fix error messages formatting
- | + librbd: ignore partial refresh error when acquiring exclusive lock
- | + librbd: potential seg fault when blacklisting an image client
- | + librbd: potential double-unwatch of watch handle upon error
- | + librbd: deadlock when replaying journal during image open
- | + librbd: improve image state machine debug log messages
- | + librbd: remove unused refresh request logic
- | + librbd: interlock image refresh and lock operations
- | + librbd: image state machine now has hooks for lock requests
- | + librbd: integrate asynchronous image rewatch state machine
- | + librbd: helper state machine for asynchronous watch recovery
- | + librbd: exclusive lock now supports reacquiring a lost lock
- | + librbd: store exclusive lock cookie instead of recalculating
- | + librbd: helper state machine to update lock cookie
- | + cls_lock: support updating the lock cookie without releasing the lock
- + Pull request 12909
- |\
- | + librbd: block concurrent in-flight object map updates for the same object
- | + librbd: new block guard helper to prevent concurrent IO to blocks
- | + librbd: convert ObjectMap to template for unit testing
- | + librbd: clean up object map update interface
- | + librbd: update in-memory object map after on-disk update committed
- + Pull request 12917
- |\
- | + rgw: Handle multiple listening addreses w/ optional ssl correctly with civetweb.
- | + rgw: s3: secure_port should override port, also apply ssl default right.
- | + rgw: Get civetweb ssl enhancement: wip-listen3 = mg_get_local_addr
- | + rgw: Document that radosgw now supports SSL.
- | + rgw: civetweb/openssl: automagic: load libssl.so and libcrypto.so by soname.
- | + rgw: civetweb/openssl: Load libssl.so and libcrypto.so by soname.
- | + rgw: cmake: remove useless civetweb include path side effect.
- + Pull request 12921
- |\
- | + mds/server: skip unwanted dn in handle_client_readdir
- + Pull request 12978
- |\
- | + PrimaryLogPG::try_lock_for_read: give up if missing
- | + ReplicatedBackend: take read locks for clone sources during recovery
- + Pull request 12997
- |\
- | + rgw_rados: add guard assert in add_io()
- | + rgw_rados: sanitize dout print in GWRados::get_obj_iterate_cb
- + Pull request 13001
- |\
- | + rgw: RGWAsyncRadosRequest drops notifier ref on cancel
- | + rgw: remove circular reference in RGWAsyncRadosRequest
- | + rgw: release RGWAioCompletionNotifier refs on destruction
- + Pull request 13025
- |\
- | + Ceph-disk to use correct user in check_journal_req
- + Pull request 13029
- |\
- | + client/Client.cc: prevent segfaulting
- + Pull request 13040
- |\
- | + tests: run fs/thrash on xfs instead of btrfs
- + Pull request 13045
- |\
- | + mon: do not send duplicated osdmap msg to not sync'ed osd
- + Pull request 13047
- |\
- | + test: Update for new error message when doing scrub with deep-scrub errors
- | + test: Add test support for deep-scrub
- | + osd: Add trigger_scrub admin socket command
- | + test: Add test for keeping deep-scrub information
- | + osd: When deep-scrub errors present upgrade regular scrubs
- + Pull request 13048
- |\
- | + selinux: Allow ceph to manage tmp files
- + Pull request 13049
- |\
- | + doc: document hostname constraints for rados bench
- + Pull request 13050
- |\
- | + tests: explicitly use centos 7.3 in distros/supported
- | + qa: fixed distros links
- | + qa/distros: centos_7.yaml -> centos.yaml
- | + qa/suites: centos_7.2.yaml -> centos_7.yaml
- | + qa/distros: add centos 7.3
- | + qa/distros: add centos 7 yaml; use that instead
- + Pull request 13062
- |\
- | + msg/simple: clear_pipe when wait() is mopping up pipes
- + Pull request 13104
- |\
- | + tasks/rbd_fio: unmap rbd devices on cleanup
- | + tasks/rbd_fio: don't use sudo unnecessarily
- + Pull request 13113
- |\
- | + mds: finish clientreplay requests before requesting active state
- + Pull request 13115
- |\
- | + mon/OSDMonitor: set last_force_op_resend on overlay pool too
- + Pull request 13117
- |\
- | + mon/OSDMonitor: only show interesting flags in health warning
- + Pull request 13118
- |\
- | + librados: Memory leaks in object_list_begin and object_list_end
- + Pull request 13119
- |\
- | + qa/tasks: add test_open_ino_errors
- | + mds: propagate error encountered during opening inode by number
- + Pull request 13120
- |\
- | + client: fix Client::handle_cap_flushsnap_ack() crash
- + Pull request 13123
- |\
- | + mon/MDSMonitor: fix iterating over mutated map
- | + mon: use clearer code structure
- + Pull request 13125
- |\
- | + cephfs: fix missing ll_get for ll_walk
- + Pull request 13126
- |\
- | + mds: fix dropping events in standby replay
- + Pull request 13128
- |\
- | + journal: don't hold future lock during assignment
- + Pull request 13129
- |\
- | + rbd: bench-write should return error if io-size >= 4G
- + Pull request 13131
- |\
- | + OSDMonitor: clear jewel+ feature bits when talking to Hammer OSD
- + Pull request 13139
- |\
- | + src/mds: fix MDSMap upgrade decoding
- | + mds: use FSMap::insert to add to standby_daemons
- + Pull request 13143
- |\
- | + radosgw/swift: clean up flush / newline behavior.
- + Pull request 13153
- |\
- | + tests: upgrade: install firefly only on Ubuntu 14.04
- + Pull request 13155
- |\
- | + rbd-mirror: avoid processing new events after stop requested
- + Pull request 13156
- |\
- | + librbd: don't remove an image w/ incompatible features
- + Pull request 13157
- |\
- | + rbd: enabling/disabling rbd feature should report missing dependency
- + Pull request 13161
- |\
- | + tests: add require_jewel_osds before upgrading last hammer node
- + Pull request 13168
- |\
- | + librbd: metadata_set API operation should not change global config setting
- + Pull request 13170
- |\
- | + tests: drop old rados singleton upgrade tests
- + Pull request 13171
- |\
- | + rgw: clear master_zonegroup when reseting RGWPeriodMap
- + Pull request 13173
- |\
- | + rgw: complete versioning enablement after sending it to meta master
- + Pull request 13175
- |\
- | + rgw_admin: read master log shards from master's current period
- | + rgw: allow getting master log shards info on specified period
- | + rgw_admin: get master's period from store's current period info
- + Pull request 13177
- |\
- | + rgw_file: add timed namespace invalidation
- + Pull request 13178
- |\
- | + rgw_rados: sanitize dout print in GWRados::get_obj_iterate_cb
- | + rgw_rados: add guard assert in add_io()
- + Pull request 13179
- |\
- | + rgw: RGWAsyncRadosRequest drops notifier ref on cancel
- | + rgw: remove circular reference in RGWAsyncRadosRequest
- | + rgw: release RGWAioCompletionNotifier refs on destruction
- + Pull request 13180
- |\
- | + rgw: fix off-by-one in RGWDataChangesLog::get_info
- + Pull request 13182
- |\
- | + radosgw-admin: check for name mistmatch in realm set
- | + radosgw-admin: relam set can use input redirection
- | + radosgw-admin: realm set should create a new realm
- + Pull request 13183
- |\
- | + build/ops: add libldap dependency for RGW
- + Pull request 13184
- + systemd: Restart Mon after 10s in case of failure
#45 Updated by Nathan Cutler over 6 years ago
rados¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 50)/50 --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-30_11:11:11-rados-wip-jewel-backports-distro-basic-smithi/
- false positive saw valgrind issues
- 764136: false positive (tcmalloc)
- known intermittent bug, succeeded on re-run http://tracker.ceph.com/issues/16236 ("racing read got wrong version")
- 764195
- known intermittent bug http://tracker.ceph.com/issues/16943 Error ENOENT: unrecognized pool '.rgw.control' (CRITICAL:root:IOError) in s3tests-test-readwrite under thrashosds
- 764231: rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml hobj-sort.yaml msgr-failures/osd-delay.yaml msgr/async.yaml rados.yaml thrashers/morepggrow.yaml workloads/rgw_snaps.yaml}
- problem with the tests, now being fixed in lfn-upgrade-{hammer,infernalis}.yaml
- 764236 "rados/singleton-nomsgr/{all/lfn-upgrade-hammer.yaml rados.yaml}" - "need more than 0 values to unpack" - appears to be a regression caused by smithfarm stupidity in PR#13161, fixed and re-pushed
- 764258 "rados/singleton-nomsgr/{all/lfn-upgrade-infernalis.yaml rados.yaml}" - "'wait_until_healthy' reached maximum tries (150) after waiting for 900 seconds" - ceph osd set require_jewel_osds does not get run - same stupid smithfarm mistake
- false positive saw valgrind issues
Re-running the first three failures:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-30_19:58:05-rados-wip-jewel-backports-distro-basic-smithi/
- false positive saw valgrind issues (tcmalloc) in the same test as before
- low-priority bug http://tracker.ceph.com/issues/18739 FAILED assert(0 == "out of order op")
tcmalloc valgrind failure¶
Re-running the "tcmalloc valgrind" test 10 times:
filter="rados/verify/{1thrash/default.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/simple.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml}" ./virtualenv/bin/teuthology-suite -k distro --priority 101 --suite rados --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi --num 10 --filter="$filter"
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-30_22:28:13-rados-wip-jewel-backports-distro-basic-smithi/
- reproduces 7 times out of 10
- not a blocker
16236 ("racing read got wrong version")¶
Succeeded on re-run.
s3tests-test-readwrite dies under thrashosds, out of order op¶
Test: rados/thrash/{0-size-min-size-overrides/3-size-2-min-size.yaml 1-pg-log-overrides/short_pg_log.yaml clusters/{fixed-2.yaml openstack.yaml} fs/xfs.yaml hobj-sort.yaml msgr-failures/osd-delay.yaml msgr/async.yaml rados.yaml thrashers/morepggrow.yaml workloads/rgw_snaps.yaml}
This test failed twice in two different ways.
Re-running 10 times:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_06:30:59-rados-wip-jewel-backports-distro-basic-smithi/
- two failed jobs
- eight successful jobs
(11:15:51 AM) owasserm: smithfarm, I looked at the failures and the s3_readwrite looks like a timeout issue (it happens after xfs injected stalls)
Conclusion: not a blocker
lfn-upgrade-{hammer,infernalis}.yaml failures¶
Re-running the last two failures with fixed PR branch (after prolonged trial-and-error):
teuthology-suite -k distro --priority 101 --suite rados --email ncutler@suse.com --ceph wip-lfn-upgrade-hammer --machine-type smithi --filter="rados/singleton-nomsgr/{all/lfn-upgrade-hammer.yaml rados.yaml},rados/singleton-nomsgr/{all/lfn-upgrade-infernalis.yaml rados.yaml}"
teuthology-suite -k distro --priority 101 --suite rados --email ncutler@suse.com --ceph wip-lfn-upgrade-hammer --machine-type vps --filter="rados/singleton-nomsgr/{all/lfn-upgrade-hammer.yaml rados.yaml},rados/singleton-nomsgr/{all/lfn-upgrade-infernalis.yaml rados.yaml}"
And again with --num 8
#46 Updated by Nathan Cutler over 6 years ago
Upgrade¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x --ceph wip-jewel-backports --machine-type vps --priority 1000
- fail http://pulpito.ceph.com/smithfarm-2017-01-30_12:04:19-upgrade:jewel-x-wip-jewel-backports-distro-basic-vps/
- "os_type: centos" without os_version causes jobs to fail on VPS, added another commit to https://github.com/ceph/ceph/pull/13050 to address this
- new bug, probably infrastructure noise http://tracker.ceph.com/issues/18733
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --ceph wip-jewel-backports --machine-type vps --priority 1000
- fail http://pulpito.ceph.com/smithfarm-2017-01-30_12:05:09-upgrade:hammer-x-wip-jewel-backports-distro-basic-vps/
- known bug http://tracker.ceph.com/issues/18089
CONCLUSION: nothing more to do here, needs a new integration branch
#47 Updated by Nathan Cutler over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-jewel-backports --machine-type vps --priority 1000
#48 Updated by Nathan Cutler over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports -k distro -m smithi -s powercycle -p 1000 --email loic@dachary.org
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-30_12:07:01-powercycle-wip-jewel-backports-distro-basic-smithi/
- unknown bug journal FileJournal::write_bl : write_fd failed: (28) No space left on device
- powercycle/osd/{clusters/3osd-1per-target.yaml fs/btrfs.yaml powercycle/default.yaml tasks/cfuse_workunit_kernel_untar_build.yaml}
- /a/smithfarm-2017-01-30_12:07:01-powercycle-wip-jewel-backports-distro-basic-smithi/765012
- unknown bug journal FileJournal::write_bl : write_fd failed: (28) No space left on device
Re-running the failed job 10 times:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_06:47:46-powercycle-wip-jewel-backports-distro-basic-smithi/
- four jobs passed
- four jobs failed
- two jobs (probably) died
Since there was just one failure in the entire suite, and that job only fails about 60% of the time, the overall run is ruled a pass
#49 Updated by Nathan Cutler over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-30_12:08:10-fs-wip-jewel-backports-distro-basic-smithi/
- "saw valgrind issues" (3 failed jobs)
- 84 successful jobs
Since the valgrind issues are apparently benign, ruled a pass
#50 Updated by Nathan Cutler over 6 years ago
rgw¶
teuthology-suite -k distro --priority 1000 --suite rgw --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-30_12:10:53-rgw-wip-jewel-backports-distro-basic-smithi/
- "saw valgrind issues" new bug http://tracker.ceph.com/issues/18744 (the daemons in the notcmalloc build are linked with libtcmalloc) (18 of 21 failed jobs)
- No summary info found for user: foo
- new bug ERROR: test suite for <module 's3tests.functional' from '/home/ubuntu/cephtest/s3-tests/s3tests/functional/__init__.py'>
- ansible: failed to update apt-cache
Re-running all the failed jobs, expecting a largish number of tcmalloc-related valgrind failures:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_12:35:14-rgw-wip-jewel-backports-distro-basic-smithi/
- "saw valgrind issues" new bug http://tracker.ceph.com/issues/18744 (the daemons in the notcmalloc build are linked with libtcmalloc) (six failed jobs)
- no other failures, ruled a pass
Re-running just "rgw_s3tests.yaml lockdep.yaml" to try to reproduce the s3tests.functional failure:
teuthology-suite -k distro --priority 1000 --suite rgw/verify --ceph wip-jewel-backports --machine-type smithi --email ncutler@suse.com --filter="tasks/rgw_s3tests.yaml validater/lockdep.yaml" --num 5
#51 Updated by Nathan Cutler over 6 years ago
rbd¶
teuthology-suite -k distro --priority 101 --suite rbd --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-01-30_12:16:19-rbd-wip-jewel-backports-distro-basic-smithi/
- newly reported, known intermittent bug http://tracker.ceph.com/issues/18731
- dead job, known bug http://tracker.ceph.com/issues/16263
failed job¶
Re-running the single failed job:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-30_19:52:39-rbd-wip-jewel-backports-distro-basic-smithi/
- newly reported, but known bug http://tracker.ceph.com/issues/18739 FAILED assert(0 == "out of order op")
Re-running it again with --num 10
:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-30_22:32:27-rbd-wip-jewel-backports-distro-basic-smithi/
- 9 pass, 1 fail
- known bug, not a blocker http://tracker.ceph.com/issues/18731
dead job¶
Re-running with --num 10
:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-30_22:56:59-rbd-wip-jewel-backports-distro-basic-smithi/
- two dead jobs (thrashosds failed to recover before timeout expired)
- eight successful jobs
conclusion¶
Ruled a pass
#52 Updated by Nathan Cutler over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel..wip-jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 10661
- |\
- | + rgw: Have a flavor of bucket deletion to bypass GC and to trigger object deletions async.
- + Pull request 11470
- |\
- | + rgw:bucket check remove multipart prefix
- + Pull request 11476
- |\
- | + rgw: RGWCoroutinesManager::run returns status of last cr
- + Pull request 11477
- |\
- | + rgw: fix for assertion in RGWMetaSyncCR
- + Pull request 11497
- |\
- | + rgw: settle /info implementation across other swift-at-root features.
- | + swift /info implementation.
- | + rgw: add support for the healthcheck feature of Swift API.
- | + rgw: add support for the crossdomain.xml resource of Swift API.
- | + rgw: fix the handling of rgw_swift_url_prefix.
- | + rgw_http_errors: add http error code for 503
- | + rgw: Allow to serve Swift off the URL root
- + Pull request 11866
- |\
- | + rgw: clean up RGWShardedOmapCRManager on early return
- + Pull request 11868
- |\
- | + rgw: store oldest mdlog period in rados
- + Pull request 11872
- |\
- | + rgw: delete entries_index in RGWFetchAllMetaCR
- + Pull request 11876
- |\
- | + rgw: fix the field 'total_time' of log entry in log show opt
- + Pull request 11990
- |\
- | + rgw: for the create_bucket api, if the input creation_time is zero, we should set it to 'now
- + Pull request 11991
- |\
- | + osd: limit omap data in push op
- + Pull request 11998
- |\
- | + FileStore::_do_fiemap: do not reference fiemap after it is freed
- | + test: add test for fiemap xfs issue when #extents > 1364
- | + FileStore:: fix fiemap issue in xfs when #extents > 1364
- + Pull request 12044
- |\
- | + rgw: fix for bucket delete racing with mdlog sync
- + Pull request 12045
- |\
- | + rgw_file: fix spurious mount entries w/Linux NFS client
- + Pull request 12079
- |\
- | + rgw: TempURL properly handles accounts created with the implicit tenant.
- + Pull request 12156
- |\
- | + rgw: look for region_map in rgw_region_root_pool
- | + rgw: region conversion respects pre-existing rgw_region_root_pool
- + Pull request 12239
- |\
- | + jewel: fixed the issue when --disable-server, compilation fails.
- + Pull request 12258
- |\
- | + rgw: add support for the prefix parameter in account listing of Swift API.
- | + rgw: optimize out ctor-copy in RGWListBuckets_ObjStore_SWIFT.
- + Pull request 12291
- |\
- | + msg/simple/Pipe: handle addr decode error
- + Pull request 12302
- |\
- | + tests: check hostname --fqdn sanity before running cmake check
- | + tests: check hostname --fqdn sanity before running make check
- + Pull request 12313
- |\
- | + rgw: only set CURLOPT_UPLOAD for PUT/POST requests
- + Pull request 12314
- |\
- | + rgw: RGWBucketSyncStatusManager uses existing async_rados
- + Pull request 12315
- |\
- | + rgw: fix missing master zone for a single zone zonegroup
- + Pull request 12316
- |\
- | + rgw: add recovery procedure for upgrade to older version of jewel
- + Pull request 12380
- |\
- | + os/ObjectStore: properly clone object map when replaying OP_COLL_MOVE_RENAME
- | + os/ObjectStore: properly clear object map when replaying OP_REMOVE
- + Pull request 12397
- |\
- | + rgw: do not abort when accept a CORS request with short origin
- + Pull request 12405
- |\
- | + install-deps.sh: unify indentation in case statement
- | + install-deps.sh: allow building on SLES systems
- | + build/ops: fix /etc/os-release parsing in install-deps.sh
- | + install-deps.sh: initial distro detection based on /etc/os-release
- + Pull request 12419
- |\
- | + rgw: omap_get_all() fixes
- | + rgw/rgw_rados: do not omap_getvals with (u64)-1 max
- + Pull request 12426
- |\
- | + rbd: --max_part and --nbds_max options for nbd map
- + Pull request 12428
- |\
- | + radosgw-admin: 'zone placement modify' doesnt require pool names
- | + radosgw-admin: add 'zonegroup placement default' command
- | + radosgw-admin: fix 'placment' typos
- | + rgw_admin: commands to manage placement targets
- | + rgw-admin: add commands to manage zonegroup placement fields
- | + rgw: use set for zonegroup placement target tags
- + Pull request 12542
- |\
- | + rgw: Replacing '+' with %20 in canonical uri for s3 v4 auth.
- + Pull request 12622
- |\
- | + rgw: log name instead of id for SystemMetaObj on failure
- | + rgw: drop unnecessary spacing in rgw zg init log
- + Pull request 12677
- |\
- | + mon: OSDMonitor: trigger an immediate propose if any newly down osd is detected during tick()
- + Pull request 12678
- |\
- | + rgw: ldap: simple_bind() should set ldap option on tldap
- + Pull request 12738
- |\
- | + rgw: use explicit flag to cancel RGWCoroutinesManager::run()
- + Pull request 12755
- |\
- | + rados: optionally support reading omap key from file
- + Pull request 12764
- |\
- | + rgw: fix decoding of creation_time and last_update.
- + Pull request 12890
- |\
- | + librbd: ensure owner lock is held before purging cache
- | + librados: blacklist_add should wait for latest OSD map
- | + librbd: prevent assertion failure when journal IO is blacklisted
- | + librbd: ignore blacklist error when releasing exclusive lock
- | + librbd: fail immediately if the exclusive lock cannot be acquired
- | + librbd: add new lock_get_owners / lock_break_lock API methods
- | + librbd: separate break lock logic into standalone state machine
- | + librbd: separate locker query into standalone state machine
- | + librbd/exclusive_lock/AcquireRequest.cc: init lock_type
- | + librbd: API methods to directly acquire and release the exclusive lock
- | + rbd-mirror: fix error messages formatting
- | + librbd: ignore partial refresh error when acquiring exclusive lock
- | + librbd: potential seg fault when blacklisting an image client
- | + librbd: potential double-unwatch of watch handle upon error
- | + librbd: deadlock when replaying journal during image open
- | + librbd: improve image state machine debug log messages
- | + librbd: remove unused refresh request logic
- | + librbd: interlock image refresh and lock operations
- | + librbd: image state machine now has hooks for lock requests
- | + librbd: integrate asynchronous image rewatch state machine
- | + librbd: helper state machine for asynchronous watch recovery
- | + librbd: exclusive lock now supports reacquiring a lost lock
- | + librbd: store exclusive lock cookie instead of recalculating
- | + librbd: helper state machine to update lock cookie
- | + cls_lock: support updating the lock cookie without releasing the lock
- + Pull request 12917
- |\
- | + rgw: Handle multiple listening addreses w/ optional ssl correctly with civetweb.
- | + rgw: s3: secure_port should override port, also apply ssl default right.
- | + rgw: Get civetweb ssl enhancement: wip-listen3 = mg_get_local_addr
- | + rgw: Document that radosgw now supports SSL.
- | + rgw: civetweb/openssl: automagic: load libssl.so and libcrypto.so by soname.
- | + rgw: civetweb/openssl: Load libssl.so and libcrypto.so by soname.
- | + rgw: cmake: remove useless civetweb include path side effect.
- + Pull request 12978
- |\
- | + PrimaryLogPG::try_lock_for_read: give up if missing
- | + ReplicatedBackend: take read locks for clone sources during recovery
- + Pull request 12997
- |\
- | + rgw_rados: add guard assert in add_io()
- | + rgw_rados: sanitize dout print in GWRados::get_obj_iterate_cb
- + Pull request 13001
- |\
- | + rgw: RGWAsyncRadosRequest drops notifier ref on cancel
- | + rgw: remove circular reference in RGWAsyncRadosRequest
- | + rgw: release RGWAioCompletionNotifier refs on destruction
- + Pull request 13025
- |\
- | + Ceph-disk to use correct user in check_journal_req
- + Pull request 13047
- |\
- | + test: Update for new error message when doing scrub with deep-scrub errors
- | + test: Add test support for deep-scrub
- | + osd: Add trigger_scrub admin socket command
- | + test: Add test for keeping deep-scrub information
- | + osd: When deep-scrub errors present upgrade regular scrubs
- + Pull request 13048
- |\
- | + selinux: Allow ceph to manage tmp files
- + Pull request 13050
- |\
- | + tests: make distros/all/centos.yaml be a symlink to centos_7.3
- | + tests: explicitly use centos 7.3 in distros/supported
- | + qa: fixed distros links
- | + qa/distros: centos_7.yaml -> centos.yaml
- | + qa/suites: centos_7.2.yaml -> centos_7.yaml
- | + qa/distros: add centos 7.3
- | + qa/distros: add centos 7 yaml; use that instead
- + Pull request 13062
- |\
- | + msg/simple: clear_pipe when wait() is mopping up pipes
- + Pull request 13106
- |\
- | + Revert use the create option during instantiation
- | + use dev option instead of dev-commit
- + Pull request 13113
- |\
- | + mds: finish clientreplay requests before requesting active state
- + Pull request 13118
- |\
- | + librados: Memory leaks in object_list_begin and object_list_end
- + Pull request 13131
- |\
- | + OSDMonitor: clear jewel+ feature bits when talking to Hammer OSD
- + Pull request 13139
- |\
- | + src/mds: fix MDSMap upgrade decoding
- | + mds: use FSMap::insert to add to standby_daemons
- + Pull request 13143
- |\
- | + radosgw/swift: clean up flush / newline behavior.
- + Pull request 13153
- |\
- | + tests: upgrade: install firefly only on Ubuntu 14.04
- + Pull request 13171
- |\
- | + rgw: clear master_zonegroup when reseting RGWPeriodMap
- + Pull request 13173
- |\
- | + rgw: complete versioning enablement after sending it to meta master
- + Pull request 13175
- |\
- | + rgw_admin: read master log shards from master's current period
- | + rgw: allow getting master log shards info on specified period
- | + rgw_admin: get master's period from store's current period info
- + Pull request 13177
- |\
- | + rgw_file: add timed namespace invalidation
- + Pull request 13178
- |\
- | + rgw_rados: sanitize dout print in GWRados::get_obj_iterate_cb
- | + rgw_rados: add guard assert in add_io()
- + Pull request 13179
- |\
- | + rgw: RGWAsyncRadosRequest drops notifier ref on cancel
- | + rgw: remove circular reference in RGWAsyncRadosRequest
- | + rgw: release RGWAioCompletionNotifier refs on destruction
- + Pull request 13180
- |\
- | + rgw: fix off-by-one in RGWDataChangesLog::get_info
- + Pull request 13182
- |\
- | + radosgw-admin: check for name mistmatch in realm set
- | + radosgw-admin: relam set can use input redirection
- | + radosgw-admin: realm set should create a new realm
- + Pull request 13183
- |\
- | + build/ops: add libldap dependency for RGW
- + Pull request 13184
- |\
- | + systemd: Restart Mon after 10s in case of failure
- + Pull request 13187
- + ceph-disk: convert none str to str before printing it
#53 Updated by Nathan Cutler over 6 years ago
rados¶
teuthology-suite -k distro --priority 101 --suite rados --subset $(expr $RANDOM % 50)/50 --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_20:05:35-rados-wip-jewel-backports-distro-basic-smithi/
- infrastructure noise clock skew
- 770350
- 770354
- 770358
- false negative http://tracker.ceph.com/issues/18744 (Leak_StillReachable in MON, with libtcmalloc frame)
- infrastructure noise SSH connection to smithi178 was lost
- infrastructure noise clock skew
Re-running failed and dead jobs:
#54 Updated by Nathan Cutler over 6 years ago
Upgrade¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x --ceph wip-jewel-backports --machine-type vps --priority 1000 --email ncutler@suse.com
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_20:13:18-upgrade:jewel-x-wip-jewel-backports-distro-basic-vps/
- timeout in radosbench
- upgrade:jewel-x/stress-split/{0-cluster/{openstack.yaml start.yaml} 1-jewel-install/jewel.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/{rbd-cls.yaml rbd-import-export.yaml readwrite.yaml snaps-few-objects.yaml} 6-next-mon/monb.yaml 7-workload/{radosbench.yaml rbd_api.yaml} 8-next-mon/monc.yaml 9-workload/{rbd-python.yaml rgw-swift.yaml snaps-many-objects.yaml} distros/ubuntu_14.04.yaml}
- smithfarm note: "I looked at this test and noticed that it installs "jewel" (i.e. latest not-yet-released) on two cluster nodes and one client node, and then upgrades just one cluster node to "-x" which is the integration branch. Shouldn't it install, say, v10.2.0 instead of latest jewel??"
- timeout in radosbench
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --ceph wip-jewel-backports --machine-type vps --priority 1000 --email ncutler@suse.com
In general, both of these runs have improved significantly but still need work:
#55 Updated by Nathan Cutler over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-jewel-backports --machine-type vps --priority 1000 --email ncutler@suse.com
#56 Updated by Nathan Cutler over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports -k distro -m smithi -s powercycle -p 1000 --email ncutler@suse.com
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_20:15:14-powercycle-wip-jewel-backports-distro-basic-smithi/
- one failure, looks similar to the previous run
Ruled a pass
#57 Updated by Nathan Cutler over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_20:16:06-fs-wip-jewel-backports-distro-basic-smithi/
- saw valgrind issues
- probable "out of order op" (just an unverified suspicion)
- cephfs failure (same as before, not a blocker)
- one dead job
Re-running the six failed jobs:
Re-running the one dead job:
- pending
#58 Updated by Nathan Cutler over 6 years ago
rgw¶
teuthology-suite -k distro --priority 101 --suite rgw --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_20:01:05-rgw-wip-jewel-backports-distro-basic-smithi/
- saw valgrind issues (18 failures) most likely http://tracker.ceph.com/issues/18744
Re-running:
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-02-01_11:09:07-rgw-wip-jewel-backports-distro-basic-smithi/
- saw valgrind issues (down to 8 failures) all checked and verified to be http://tracker.ceph.com/issues/18744
Conclusion: since all 8 failures are libtcmalloc-related, Orit says we can merge all the RGW backport PRs.
#59 Updated by Nathan Cutler over 6 years ago
rbd¶
teuthology-suite -k distro --priority 1000 --suite rbd --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.front.sepia.ceph.com:80/smithfarm-2017-01-31_20:17:40-rbd-wip-jewel-backports-distro-basic-smithi/
- known, low-priority bug http://tracker.ceph.com/issues/18739 FAILED assert(0 == "out of order op")
Ruled a pass
#60 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#61 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#62 Updated by Nathan Cutler over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel..wip-jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 12044
- |\
- | + rgw: fix for bucket delete racing with mdlog sync
- + Pull request 12079
- |\
- | + rgw: TempURL properly handles accounts created with the implicit tenant.
- + Pull request 12239
- |\
- | + jewel: fixed the issue when --disable-server, compilation fails.
- + Pull request 12380
- |\
- | + os/ObjectStore: properly clone object map when replaying OP_COLL_MOVE_RENAME
- | + os/ObjectStore: properly clear object map when replaying OP_REMOVE
- + Pull request 12490
- |\
- | + rgw: add 'rgw log http headers' gloss to config-ref.rst
- | + use std::map
- | + rgw: add rgw_log_http_headers option
- + Pull request 12729
- |\
- | + jewel: fix compile error for dencode test case when --with-radosgw=no
- | + jewel: fixed compile error when --with-radosgw=no
- + Pull request 12754
- |\
- | + rbd-nbd: support partition for rbd-nbd mapped raw block device.
- + Pull request 12917
- |\
- | + rgw: Handle multiple listening addreses w/ optional ssl correctly with civetweb.
- | + rgw: s3: secure_port should override port, also apply ssl default right.
- | + rgw: Get civetweb ssl enhancement: wip-listen3 = mg_get_local_addr
- | + rgw: Document that radosgw now supports SSL.
- | + rgw: civetweb/openssl: automagic: load libssl.so and libcrypto.so by soname.
- | + rgw: civetweb/openssl: Load libssl.so and libcrypto.so by soname.
- | + rgw: cmake: remove useless civetweb include path side effect.
- + Pull request 13004
- |\
- | + rgw: RGWCloneMetaLogCoroutine uses RGWMetadataLogInfoCompletion
- | + rgw: expose completion for RGWMetadataLog::get_info_async()
- | + rgw: RGWMetaSyncShardCR drops stack refs on destruction
- | + rgw: librados aio wait_for_safe, not wait_for_complete
- + Pull request 13048
- |\
- | + selinux: Allow ceph to manage tmp files
- + Pull request 13050
- |\
- | + tests: make distros/all/centos.yaml be a symlink to centos_7.3
- | + tests: explicitly use centos 7.3 in distros/supported
- | + qa: fixed distros links
- | + qa/distros: centos_7.yaml -> centos.yaml
- | + qa/suites: centos_7.2.yaml -> centos_7.yaml
- | + qa/distros: add centos 7.3
- | + qa/distros: add centos 7 yaml; use that instead
- + Pull request 13058
- |\
- | + systemd: Restart Mon after 10s in case of failure
- + Pull request 13113
- |\
- | + mds: finish clientreplay requests before requesting active state
- + Pull request 13130
- |\
- | + librbd: allow to open an image without opening parent image
- + Pull request 13131
- |\
- | + OSDMonitor: clear jewel+ feature bits when talking to Hammer OSD
- + Pull request 13139
- |\
- | + src/mds: fix MDSMap upgrade decoding
- | + mds: use FSMap::insert to add to standby_daemons
- + Pull request 13146
- |\
- | + test: Update for new error message when doing scrub with deep-scrub errors
- | + osd: Add trigger_scrub admin socket command
- | + test: Add test for keeping deep-scrub information
- | + osd: When deep-scrub errors present upgrade regular scrubs
- | + tasks/scrub_test.py: Make test deterministic by updating digests
- | + repair_test, scrub_test: Fix whitelists for scrub changes
- | + scrub_test: Fix for list-inconsistent-obj output changes
- | + doc, test: Add schemas for list-inconsistent-* rados command output
- | + test: Update testing for new list-inconsistent-obj output
- | + rados, osd: Improve attrs output of list-inconsistent-obj
- | + osd: Fix logging to help with diagnostics
- | + test: Fix use of wait_for_clean()
- | + common: Change cleanbin() to use base64 encoding, update ceph-objectstore-tool
- | + common: Move cleanbin() function to common/util.cc
- | + test: Add test support for deep-scrub
- | + common: Fix indentation
- | + osd: Handle corrupt attributes in get_object_context()
- | + ReplicatedPG::failed_push: release read lock on failure
- | + test.sh: Make check for flags more robust
- | + test: Remove extra objectstore_tool call which causes a recovery
- | + test: Handle object removals in a non-racey way
- | + osd: Fix hang on unfound object after mark_unfound_lost is done
- | + osd: Handle recovery read errors
- | + osd: Fix log messages
- | + osd: CLEANUP: Remove unused pending_read member
- | + test/osd-scrub-repair.sh: Use test case specific object names to help with diagnostics
- + Pull request 13183
- |\
- | + build/ops: add libldap dependency for RGW
- + Pull request 13184
- |\
- | + systemd: Restart Mon after 10s in case of failure
- + Pull request 13212
- |\
- | + test/osd: add test for fast mark down functionality
- | + msg/async: implement ECONNREFUSED detection
- | + messages/MOSDFailure.h: distinguish between timeout and immediate failure
- | + OSD: Implement ms_handle_refused
- | + msg/simple: add ms_handle_refused callback
- | + AsyncConnection: fix delay state using dispatch_queue
- | + AsyncConnection: need to prepare message when features mismatch
- | + AsyncConnection: continue to read when meeting EINTR
- | + AsyncConnection: release dispatch throttle with fast dispatch message
- | + DispatchQueue: remove pipe words
- | + DispatchQueue: add name to separte different instance
- | + AsyncConnection: add DispathQueue throttle
- | + AsyncConnection: change all exception deliver to DispatchQueue
- | + AsyncConnection: make local message deliver via DispatchQueue
- | + AsyncMessenger: introduce DispatchQueue to separate nonfast message
- | + DispatchQueue: move dispatch_throtter from SimpleMessenger to DispatchQueue
- | + DispatchQueue: Move from msg/simple to msg
- + Pull request 13214
- |\
- | + OSD: allow client throttler to be adjusted on-fly, without restart
- + Pull request 13222
- |\
- | + qa/suites/upgrade/hammer-x: wrap thrash and workloads
- + Pull request 13232
- |\
- | + osd: Increase priority for inactive PGs backfill
- + | Pull request 13233
- |\ \
- | + | librbd: async method to check journal tag owner
- | + | rbd-mirror: check image mirroring state when bootstrapping
- | + | rbd-mirror: async request to test if image is primary
- | + | rbd-mirror: hold owner lock when testing if lock owner
- | /
- + | Pull request 13240
- |\ \
- | + | tests: fix regression in qa/tasks/ceph_master.py
- | + | tests: ignore bogus ceph-objectstore-tool error in ceph_manager
- | /
- + | Pull request 13244
- |\ \
- | + | osdc: cache should ignore error bhs during trim
- | /
- + | Pull request 13254
- |\ \
- | + | radosstriper : protect aio_write API from calls with 0 bytes
- | /
- + | Pull request 13255
- |\ \
- | |/
- | + osd: do not send ENXIO on misdirected op by default
- + Pull request 13261
- |\
- | + mon/OSDMonitor: make 'osd crush move ...' work on osds
- + Pull request 13273
- |\
- | + rgw: add check for update return value
- | + rgw: we need to reinit the zonegroup after assignment to avoid invalid cct and store
- | + rgw: fix init_zg_from_period when default zone is not set as default
- + Pull request 13276
- |\
- | + rgw: be aware abount tenants on cls_user_bucket -> rgw_bucket conversion.
- + Pull request 13341
- + Backport bucket reshard to jewel.
- + rgw_admin: add a few admin commands to the usage
- + rgw_admin: add bi purge command
- + rgw: bucket resharding, adjust logging
- + cls/rgw: bi_list() fix is_truncated returned param
- + rgw_admin: require --yes-i-really-mean-it for bucket reshard
- + rgw_admin: better bucket reshard logging
- + rgw: limit bucket reshard num shards to max possible
- + rgw_admin: fix bi list command
- + rgw_admin: use aio operations for bucket resharding
- + rgw: bucket reshard updates stats
- + cls/rgw: add bucket_update_stats method
- + rgw_admin: reshard also links to new bucket instance
- + rgw: rgw_link_bucket, use correct bucket structure for entry point
- + radosgw-admin: bucket reshard needs --num-shards to be specified
- + cls/rgw: fix bi_list objclass command
- + rgw_admin: bucket rehsrading, initial work
- + rgw: utilities to support raw bucket index operations
- + rgw: use bucket_info.bucket_id instead of marker where needed
- + cls/rgw: utilities to support raw bucket index operations
#63 Updated by Nathan Cutler over 6 years ago
rados¶
teuthology-suite -k distro --priority 101 --suite rados --subset $(expr $RANDOM % 50)/50 --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-02-13_20:36:56-rados-wip-jewel-backports-distro-basic-smithi/
python ../fail.py $run fail- **
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=86f3d99a9d10afc8be7a72b4648550687dcf6cf1 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/mon/crush_ops.sh'- rados/monthrash/{ceph/ceph.yaml clusters/3-mons.yaml fs/xfs.yaml msgr-failures/mon-delay.yaml msgr/simple.yaml rados.yaml thrashers/many.yaml workloads/rados_mon_workunits.yaml}
- rados/monthrash/{ceph/ceph.yaml clusters/9-mons.yaml fs/xfs.yaml msgr-failures/few.yaml msgr/async.yaml rados.yaml thrashers/one.yaml workloads/rados_mon_workunits.yaml}
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&ref=v0.80.8
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=9d446bd416c52cd785ccf048ca67737ceafcdd7f
- saw valgrind issues
#64 Updated by Nathan Cutler over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports -k distro -m smithi -s powercycle -p 1000 --email ncutler@suse.com
#65 Updated by Nathan Cutler over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
#66 Updated by Nathan Cutler over 6 years ago
rgw¶
teuthology-suite -k distro --priority 101 --suite rgw --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
#67 Updated by Nathan Cutler over 6 years ago
rbd¶
teuthology-suite -k distro --priority 1000 --suite rbd --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-02-13_22:42:25-rbd-wip-jewel-backports-distro-basic-smithi/
- 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster cluster2 -i 1'
- dead job is qemu_xfstest
Re-running failed jobs
#68 Updated by Loïc Dachary over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel..wip-jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 12239
- |\
- | + jewel: fixed the issue when --disable-server, compilation fails.
- + Pull request 12380
- |\
- | + os/ObjectStore: properly clone object map when replaying OP_COLL_MOVE_RENAME
- | + os/ObjectStore: properly clear object map when replaying OP_REMOVE
- + Pull request 12729
- |\
- | + jewel: fix compile error for dencode test case when --with-radosgw=no
- | + jewel: fixed compile error when --with-radosgw=no
- + Pull request 12754
- |\
- | + rbd-nbd: support partition for rbd-nbd mapped raw block device.
- + Pull request 12917
- |\
- | + rgw: Handle multiple listening addreses w/ optional ssl correctly with civetweb.
- | + rgw: s3: secure_port should override port, also apply ssl default right.
- | + rgw: Get civetweb ssl enhancement: wip-listen3 = mg_get_local_addr
- | + rgw: Document that radosgw now supports SSL.
- | + rgw: civetweb/openssl: automagic: load libssl.so and libcrypto.so by soname.
- | + rgw: civetweb/openssl: Load libssl.so and libcrypto.so by soname.
- | + rgw: cmake: remove useless civetweb include path side effect.
- + Pull request 13004
- |\
- | + rgw: RGWCloneMetaLogCoroutine uses RGWMetadataLogInfoCompletion
- | + rgw: expose completion for RGWMetadataLog::get_info_async()
- | + rgw: RGWMetaSyncShardCR drops stack refs on destruction
- | + rgw: librados aio wait_for_safe, not wait_for_complete
- + Pull request 13048
- |\
- | + selinux: Allow ceph to manage tmp files
- + Pull request 13050
- |\
- | + tests: make distros/all/centos.yaml be a symlink to centos_7.3
- | + tests: explicitly use centos 7.3 in distros/supported
- | + qa: fixed distros links
- | + qa/distros: centos_7.yaml -> centos.yaml
- | + qa/suites: centos_7.2.yaml -> centos_7.yaml
- | + qa/distros: add centos 7.3
- | + qa/distros: add centos 7 yaml; use that instead
- + Pull request 13058
- |\
- | + systemd: Restart Mon after 10s in case of failure
- + Pull request 13130
- |\
- | + librbd: allow to open an image without opening parent image
- + Pull request 13131
- |\
- | + OSDMonitor: clear jewel+ feature bits when talking to Hammer OSD
- + Pull request 13146
- |\
- | + test: Update for new error message when doing scrub with deep-scrub errors
- | + osd: Add trigger_scrub admin socket command
- | + test: Add test for keeping deep-scrub information
- | + osd: When deep-scrub errors present upgrade regular scrubs
- | + tasks/scrub_test.py: Make test deterministic by updating digests
- | + repair_test, scrub_test: Fix whitelists for scrub changes
- | + scrub_test: Fix for list-inconsistent-obj output changes
- | + doc, test: Add schemas for list-inconsistent-* rados command output
- | + test: Update testing for new list-inconsistent-obj output
- | + rados, osd: Improve attrs output of list-inconsistent-obj
- | + osd: Fix logging to help with diagnostics
- | + test: Fix use of wait_for_clean()
- | + common: Change cleanbin() to use base64 encoding, update ceph-objectstore-tool
- | + common: Move cleanbin() function to common/util.cc
- | + test: Add test support for deep-scrub
- | + common: Fix indentation
- | + osd: Handle corrupt attributes in get_object_context()
- | + ReplicatedPG::failed_push: release read lock on failure
- | + test.sh: Make check for flags more robust
- | + test: Remove extra objectstore_tool call which causes a recovery
- | + test: Handle object removals in a non-racey way
- | + osd: Fix hang on unfound object after mark_unfound_lost is done
- | + osd: Handle recovery read errors
- | + osd: Fix log messages
- | + osd: CLEANUP: Remove unused pending_read member
- | + test/osd-scrub-repair.sh: Use test case specific object names to help with diagnostics
- + Pull request 13183
- |\
- | + build/ops: add libldap dependency for RGW
- + Pull request 13184
- |\
- | + systemd: Restart Mon after 10s in case of failure
- + Pull request 13212
- |\
- | + test/osd: add test for fast mark down functionality
- | + msg/async: implement ECONNREFUSED detection
- | + messages/MOSDFailure.h: distinguish between timeout and immediate failure
- | + OSD: Implement ms_handle_refused
- | + msg/simple: add ms_handle_refused callback
- | + AsyncConnection: fix delay state using dispatch_queue
- | + AsyncConnection: need to prepare message when features mismatch
- | + AsyncConnection: continue to read when meeting EINTR
- | + AsyncConnection: release dispatch throttle with fast dispatch message
- | + DispatchQueue: remove pipe words
- | + DispatchQueue: add name to separte different instance
- | + AsyncConnection: add DispathQueue throttle
- | + AsyncConnection: change all exception deliver to DispatchQueue
- | + AsyncConnection: make local message deliver via DispatchQueue
- | + AsyncMessenger: introduce DispatchQueue to separate nonfast message
- | + DispatchQueue: move dispatch_throtter from SimpleMessenger to DispatchQueue
- | + DispatchQueue: Move from msg/simple to msg
- + Pull request 13214
- |\
- | + OSD: allow client throttler to be adjusted on-fly, without restart
- + Pull request 13222
- |\
- | + qa/suites/upgrade/hammer-x: wrap thrash and workloads
- + Pull request 13232
- |\
- | + osd: Increase priority for inactive PGs backfill
- + | Pull request 13233
- |\ \
- | + | librbd: async method to check journal tag owner
- | + | rbd-mirror: check image mirroring state when bootstrapping
- | + | rbd-mirror: async request to test if image is primary
- | + | rbd-mirror: hold owner lock when testing if lock owner
- | /
- + | Pull request 13240
- |\ \
- | + | tests: fix regression in qa/tasks/ceph_master.py
- | + | tests: ignore bogus ceph-objectstore-tool error in ceph_manager
- | /
- + | Pull request 13244
- |\ \
- | + | osdc: cache should ignore error bhs during trim
- | /
- + | Pull request 13254
- |\ \
- | + | radosstriper : protect aio_write API from calls with 0 bytes
- | /
- + | Pull request 13255
- |\ \
- | |/
- | + osd: do not send ENXIO on misdirected op by default
- + Pull request 13273
- |\
- | + rgw: add check for update return value
- | + rgw: we need to reinit the zonegroup after assignment to avoid invalid cct and store
- | + rgw: fix init_zg_from_period when default zone is not set as default
- + Pull request 13276
- |\
- | + rgw: be aware abount tenants on cls_user_bucket -> rgw_bucket conversion.
- + Pull request 13341
- + Backport bucket reshard to jewel.
- + rgw_admin: add a few admin commands to the usage
- + rgw_admin: add bi purge command
- + rgw: bucket resharding, adjust logging
- + cls/rgw: bi_list() fix is_truncated returned param
- + rgw_admin: require --yes-i-really-mean-it for bucket reshard
- + rgw_admin: better bucket reshard logging
- + rgw: limit bucket reshard num shards to max possible
- + rgw_admin: fix bi list command
- + rgw_admin: use aio operations for bucket resharding
- + rgw: bucket reshard updates stats
- + cls/rgw: add bucket_update_stats method
- + rgw_admin: reshard also links to new bucket instance
- + rgw: rgw_link_bucket, use correct bucket structure for entry point
- + radosgw-admin: bucket reshard needs --num-shards to be specified
- + cls/rgw: fix bi_list objclass command
- + rgw_admin: bucket rehsrading, initial work
- + rgw: utilities to support raw bucket index operations
- + rgw: use bucket_info.bucket_id instead of marker where needed
- + cls/rgw: utilities to support raw bucket index operations
#69 Updated by Loïc Dachary over 6 years ago
rados¶
teuthology-suite -k distro --priority 1000 --suite rados --subset $(expr $RANDOM % 50)/50 --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-02-16_09:52:21-rados-wip-jewel-backports-distro-basic-smithi
- **
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&ref=v0.80.8
- "2017-02-16 13:16:40.828313 osd.0 172.21.15.33:6804/12748 8 : cluster [WRN] map e388 wrongly marked me down" in cluster log
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=9d446bd416c52cd785ccf048ca67737ceafcdd7f
- "2017-02-16 13:35:11.781460 osd.3 172.21.15.202:6800/22006 1 : cluster [WRN] map e200 wrongly marked me down" in cluster log
- saw valgrind issues
Re-running failed jobs
- fail http://pulpito.ceph.com/loic-2017-02-20_07:43:57-rados-wip-jewel-backports-distro-basic-smithi
- known bug assert len(unclean) == num_unclean in dump_stuck.py in rados suite
- "2017-02-20 08:15:50.988177 osd.0 172.21.15.126:6800/675602 1 : cluster [WRN] map e420 wrongly marked me down" in cluster log
- "2017-02-20 08:09:21.196329 osd.1 172.21.15.179:6806/27276 1 : cluster [WRN] map e14 wrongly marked me down" in cluster log
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=9d446bd416c52cd785ccf048ca67737ceafcdd7f
- "2017-02-20 08:27:05.421722 osd.0 172.21.15.110:6808/29569 52 : cluster [WRN] map e602 wrongly marked me down" in cluster log
Re-running on the failed jobs on jewel to assert a regression
#70 Updated by Loïc Dachary over 6 years ago
Upgrade jewel point-to-point-x¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x/point-to-point-x --ceph wip-jewel-backports --machine-type vps --priority 1000
#71 Updated by Loïc Dachary over 6 years ago
Upgrade hammer-x¶
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --ceph wip-jewel-backports --machine-type vps --priority 1000
- fail http://pulpito.ceph.com/loic-2017-02-16_09:55:58-upgrade:hammer-x-wip-jewel-backports-distro-basic-vps
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&ref=firefly
- Failed to fetch package version from https://shaman.ceph.com/api/search/?status=ready&project=ceph&flavor=default&distros=ubuntu%2F14.04%2Fx86_64&sha1=95cefea9fd9ab740263bf8bb4796fd864d9afe2b
- "sudo TESTDIR=/home/ubuntu/cephtest bash -c 'ceph osd set require_jewel_osds'"
- upgrade:hammer-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/centos_7.3.yaml}
- upgrade:hammer-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/ubuntu_14.04.yaml}
- 'sudo TESTDIR=/home/ubuntu/cephtest bash -c \'sudo ceph osd erasure-code-profile set profile-shec k=2 m=1 c=1 plugin=shec 2>&1 | grep "unsupported by"\''
- upgrade:hammer-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-no-shec.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/ubuntu_14.04.yaml}
- upgrade:hammer-x/stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-no-shec.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/centos_7.3.yaml}
Re-running
#72 Updated by Loïc Dachary over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-jewel-backports --machine-type vps --priority 1000
#73 Updated by Loïc Dachary over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports -k distro -m smithi -s powercycle -p 1000 --email loic@dachary.org
- fail http://pulpito.ceph.com/loic-2017-02-16_09:58:01-powercycle-wip-jewel-backports-distro-basic-smithi
- "sudo yum -y install '' ceph-radosgw"
- "2017-02-16 16:43:03.101694 osd.0 172.21.15.4:6800/22904 24 : cluster [ERR] 1.0 deep-scrub 1 errors" in cluster log
#74 Updated by Loïc Dachary over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-02-16_09:58:57-fs-wip-jewel-backports-distro-basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.0/client.0/tmp && cd -- /home/ubuntu/cephtest/mnt.0/client.0/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=52902891f10862c107758b1e7f0ed67edd486a89 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster ceph" CEPH_ID="0" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.client.0 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.client.0/qa/workunits/libcephfs-java/test.sh'
- 'mkdir
Re-running failed jobs
#75 Updated by Loïc Dachary over 6 years ago
rgw¶
teuthology-suite -k distro --priority 1000 --suite rgw --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com/loic-2017-02-16_10:00:14-rgw-wip-jewel-backports-distro-basic-smithi
- "sudo yum -y install '' ceph-radosgw"
- 'sudo apt-get update'
- reached maximum tries (50) after waiting for 300 seconds
- No summary info found for user: foo
Re-running failed jobs
#76 Updated by Loïc Dachary over 6 years ago
rbd¶
teuthology-suite -k distro --priority 1000 --suite rbd --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- running http://pulpito.ceph.com/loic-2017-02-16_10:01:26-rbd-wip-jewel-backports-distro-basic-smithi
- 'mkdir
p -/home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && cd -- /home/ubuntu/cephtest/mnt.cluster1.mirror/client.mirror/tmp && CEPH_CLI_TEST_DUP_COMMAND=1 CEPH_REF=52902891f10862c107758b1e7f0ed67edd486a89 TESTDIR="/home/ubuntu/cephtest" CEPH_ARGS="--cluster cluster1" CEPH_ID="mirror" PATH=$PATH:/usr/sbin CEPH_BASE=/home/ubuntu/cephtest/clone.cluster1.client.mirror CEPH_ARGS=\'\' RBD_MIRROR_USE_RBD_MIRROR=1 RBD_MIRROR_USE_EXISTING_CLUSTER=1 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 3h /home/ubuntu/cephtest/clone.cluster1.client.mirror/qa/workunits/rbd/rbd_mirror.sh'
- 'mkdir
Re-running failed jobs
#77 Updated by Loïc Dachary over 6 years ago
git --no-pager log --format='%H %s' --graph ceph/jewel..wip-jewel-backports | perl -p -e 's/"/ /g; if (/\w+\s+Merge pull request #(\d+)/) { s|\w+\s+Merge pull request #(\d+).*|"Pull request $1":https://github.com/ceph/ceph/pull/$1|; } else { s|(\w+)\s+(.*)|"$2":https://github.com/ceph/ceph/commit/$1|; } s/\*/+/; s/^/* /;'
- + Pull request 12917
- |\
- | + rgw: Handle multiple listening addreses w/ optional ssl correctly with civetweb.
- | + rgw: s3: secure_port should override port, also apply ssl default right.
- | + rgw: Get civetweb ssl enhancement: wip-listen3 = mg_get_local_addr
- | + rgw: Document that radosgw now supports SSL.
- | + rgw: civetweb/openssl: automagic: load libssl.so and libcrypto.so by soname.
- | + rgw: civetweb/openssl: Load libssl.so and libcrypto.so by soname.
- | + rgw: cmake: remove useless civetweb include path side effect.
- + Pull request 13048
- |\
- | + selinux: Allow ceph to manage tmp files
- + Pull request 13050
- |\
- | + tests: make distros/all/centos.yaml be a symlink to centos_7.3
- | + tests: explicitly use centos 7.3 in distros/supported
- | + qa: fixed distros links
- | + qa/distros: centos_7.yaml -> centos.yaml
- | + qa/suites: centos_7.2.yaml -> centos_7.yaml
- | + qa/distros: add centos 7.3
- | + qa/distros: add centos 7 yaml; use that instead
- + Pull request 13058
- |\
- | + systemd: Restart Mon after 10s in case of failure
- + Pull request 13131
- |\
- | + OSDMonitor: clear jewel+ feature bits when talking to Hammer OSD
- + Pull request 13184
- |\
- | + systemd: Restart Mon after 10s in case of failure
- + Pull request 13212
- |\
- | + test/osd: add test for fast mark down functionality
- | + msg/async: implement ECONNREFUSED detection
- | + messages/MOSDFailure.h: distinguish between timeout and immediate failure
- | + OSD: Implement ms_handle_refused
- | + msg/simple: add ms_handle_refused callback
- | + AsyncConnection: fix delay state using dispatch_queue
- | + AsyncConnection: need to prepare message when features mismatch
- | + AsyncConnection: continue to read when meeting EINTR
- | + AsyncConnection: release dispatch throttle with fast dispatch message
- | + DispatchQueue: remove pipe words
- | + DispatchQueue: add name to separte different instance
- | + AsyncConnection: add DispathQueue throttle
- | + AsyncConnection: change all exception deliver to DispatchQueue
- | + AsyncConnection: make local message deliver via DispatchQueue
- | + AsyncMessenger: introduce DispatchQueue to separate nonfast message
- | + DispatchQueue: move dispatch_throtter from SimpleMessenger to DispatchQueue
- | + DispatchQueue: Move from msg/simple to msg
- + Pull request 13214
- |\
- | + OSD: allow client throttler to be adjusted on-fly, without restart
- + Pull request 13232
- |\
- | + osd: Increase priority for inactive PGs backfill
- + | Pull request 13240
- |\ \
- | + | tests: fix regression in qa/tasks/ceph_master.py
- | + | tests: ignore bogus ceph-objectstore-tool error in ceph_manager
- | /
- + | Pull request 13244
- |\ \
- | + | osdc: cache should ignore error bhs during trim
- | /
- + | Pull request 13254
- |\ \
- | + | radosstriper : protect aio_write API from calls with 0 bytes
- | /
- + | Pull request 13255
- |\ \
- | |/
- | + osd: do not send ENXIO on misdirected op by default
- + Pull request 13450
- |\
- | + msg/simple/Pipe: support IPv6 QoS.
- + Pull request 13459
- |\
- | + mds: fix incorrect assertion in Server::_dir_is_nonempty()
- + Pull request 13477
- |\
- | + ceph-osd: --flush-journal: sporadic segfaults on exit
- + Pull request 13489
- |\
- | + ceph-disk: Fix getting wrong group name when --setgroup in bluestore
- + Pull request 13496
- + ceph-disk: change get_dmcrypt_key test to support different cluster name
- + ceph-disk: Adding cluster name support for dmcrypt
#78 Updated by Nathan Cutler over 6 years ago
fs¶
teuthology-suite -k distro --priority 1000 --suite fs --email loic@dachary.org --ceph wip-jewel-backports --machine-type smithi
- pass http://pulpito.ceph.com:80/smithfarm-2017-02-20_21:11:08-fs-wip-jewel-backports-distro-basic-smithi/
NOTE: merge https://github.com/ceph/ceph/pull/13459 when/if this run passes, and then ask John to approve 10.2.6 DONE
#79 Updated by Nathan Cutler over 6 years ago
rados¶
teuthology-suite -k distro --priority 101 --suite rados --subset $(expr $RANDOM % 50)/50 --email ncutler@suse.com --ceph wip-jewel-backports --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-02-20_21:18:05-rados-wip-jewel-backports-distro-basic-smithi/
- one failure is an instance of http://tracker.ceph.com/issues/18089 and can be ignored
- 840283 (rados/singleton/{all/osd-recovery.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/random.yaml rados.yaml}): runs for 8 hours, no sign of stopping
- 840382 (rados/singleton/{all/dump-stuck.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/random.yaml rados.yaml}): http://tracker.ceph.com/issues/17366
- one transient valgrind failure
Re-running the three other failures:
filter="rados/singleton/{all/osd-recovery.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/random.yaml rados.yaml},rados/singleton/{all/dump-stuck.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/random.yaml rados.yaml},rados/verify/{1thrash/none.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/random.yaml rados.yaml tasks/rados_cls_all.yaml validater/valgrind.yaml}"
- running http://pulpito.ceph.com:80/smithfarm-2017-02-21_14:15:43-rados-wip-jewel-backports-distro-basic-smithi/
- 843306 (rados/singleton/{all/dump-stuck.yaml fs/xfs.yaml msgr-failures/many.yaml msgr/random.yaml rados.yaml}): http://tracker.ceph.com/issues/17366
#80 Updated by Nathan Cutler over 6 years ago
powercycle¶
teuthology-suite -v -c wip-jewel-backports -k distro -m smithi -s powercycle -p 101 --email ncutler@suse.com
#81 Updated by Nathan Cutler over 6 years ago
Upgrade jewel point-to-point-x¶
teuthology-suite -k distro --verbose --suite upgrade/jewel-x/point-to-point-x --ceph wip-jewel-backports --machine-type vps --priority 101 --email ncutler@suse.com
Re-running:
#82 Updated by Nathan Cutler over 6 years ago
Upgrade hammer-x¶
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --ceph wip-jewel-backports --machine-type vps --priority 101 --email ncutler@suse.com
- fail http://pulpito.ceph.com:80/smithfarm-2017-02-20_21:28:11-upgrade:hammer-x-wip-jewel-backports-distro-basic-vps/
- two cases of http://tracker.ceph.com/issues/18089
- one mysterious failure (thrashosds/ceph_manager.py raises exception "Exception: ceph-objectstore-tool: exp list-pgs failure with status 1")
Re-running the list-pgs failure:
filter="stress-split-erasure-code/{0-cluster/{openstack.yaml start.yaml} 0-tz-eastern.yaml 1-hammer-install/hammer.yaml 2-partial-upgrade/firsthalf.yaml 3-thrash/default.yaml 4-mon/mona.yaml 5-workload/ec-rados-default.yaml 6-next-mon/monb.yaml 8-finish-upgrade/last-osds-and-monc.yaml 9-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml distros/ubuntu_14.04.yaml}"
teuthology-suite -k distro --verbose --suite upgrade/hammer-x --ceph wip-jewel-backports --machine-type vps --priority 101 --email ncutler@suse.com --filter="$filter"
#83 Updated by Nathan Cutler over 6 years ago
ceph-disk¶
teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-jewel-backports --machine-type vps --priority 101 --email ncutler@suse.com
- fail http://pulpito.ceph.com:80/smithfarm-2017-02-20_21:30:56-ceph-disk-wip-jewel-backports-distro-basic-vps/
- a bunch of tests fail with
Exception: timeout waiting for osd $SOME_UUID to be up
- a bunch of tests fail with
Because of one of :
#84 Updated by Nathan Cutler over 6 years ago
Release blockers to be merged urgently upon obtaining green rados/powercycle/upgrade:
- https://github.com/ceph/ceph/pull/13050
- https://github.com/ceph/ceph/pull/13131
- https://github.com/ceph/ceph/pull/13255
After these are merged, ask Josh to approve 10.2.6
#85 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#86 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#87 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#88 Updated by Nathan Cutler over 6 years ago
RADOS ON PR#13131 AND PR#13255¶
teuthology-suite -k distro --priority 101 --suite rados --subset $(expr $RANDOM % 50)/50 --email ncutler@suse.com --ceph pr-13131 --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-02-21_20:38:05-rados-pr-13131-distro-basic-smithi/
- one failure is http://tracker.ceph.com/issues/18089 and can be ignored
- three jobs are rados/thrash-erasure-code-isa jobs which cannot be scheduled because they want CentOS 7.2 smithis (of which there are none) - fix would be to rebase the branch to pick up PR#13050, and re-run
teuthology-suite -k distro --priority 101 --suite rados --subset $(expr $RANDOM % 50)/50 --email ncutler@suse.com --ceph pr-13255 --machine-type smithi
- fail http://pulpito.ceph.com:80/smithfarm-2017-02-21_20:43:35-rados-pr-13255-distro-basic-smithi/
- one failure is http://tracker.ceph.com/issues/18089 and can be ignored
- three jobs cannot be scheduled because they want CentOS 7.2 smithis which do not exist (missing PR#13050 in this branch)
- one instance of "api_misc: [ FAILED ] LibRadosMiscConnectFailure.ConnectFailure" in rados/verify/{1thrash/none.yaml clusters/{fixed-2.yaml openstack.yaml} fs/btrfs.yaml msgr-failures/few.yaml msgr/async.yaml rados.yaml tasks/rados_api_tests.yaml validater/lockdep.yaml}
Re-running four failed jobs (three are btrfs, one is xfs):
The one failure is in the same test, and looks very similar:
2017-02-22T15:26:29.695 INFO:tasks.workunit.client.0.smithi114.stdout: api_misc: test/librados/misc.cc:71: Failure 2017-02-22T15:26:29.695 INFO:tasks.workunit.client.0.smithi114.stdout: api_misc: Expected: (0) != (rados_connect(cluster)), actual: 0 vs 0 2017-02-22T15:26:29.695 INFO:tasks.workunit.client.0.smithi114.stdout: api_misc: [ FAILED ] LibRadosMiscConnectFailure.ConnectFailure (43 ms)
#89 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#90 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#91 Updated by Nathan Cutler over 6 years ago
- Description updated (diff)
#92 Updated by Yuri Weinstein over 6 years ago
QE VALIDATION (STARTED 2/23/17)¶
(Note: PASSED / FAILED - indicates "TEST IS IN PROGRESS")
re-runs command lines and filters are captured in http://pad.ceph.com/p/hammer_v10.2.6_QE_validation_notes
command line CEPH_QA_MAIL="ceph-qa@ceph.com"; MACHINE_NAME=smithi; CEPH_BRANCH=jewel; SHA1=d9eaab456ff45ae88e83bd633f0c4efb5902bf07 ; teuthology-suite -v --ceph-repo https://github.com/ceph/ceph.git --suite-repo https://github.com/ceph/ceph.git -c $CEPH_BRANCH -S $SHA1 -m $MACHINE_NAME -s rados --subset 35/50 -k distro -p 100 -e $CEPH_QA_MAIL --suite-branch jewel --dry-run
teuthology-suite -v -c $CEPH_BRANCH -S $SHA1 -m $MACHINE_NAME -r $RERUN --suite-repo https://github.com/ceph/ceph.git --ceph-repo https://github.com/ceph/ceph.git --suite-branch jewel -p 90 -R fail,dead,running
Suite | Runs/Reruns | Notes/Issues |
rados | http://pulpito.ceph.com/yuriw-2017-02-23_17:29:38-rados-jewel-distro-basic-smithi/ | PASSED one job #18089 Josh approved, failed job removed https://github.com/ceph/ceph/pull/13705 |
http://pulpito.ceph.com/yuriw-2017-02-24_16:57:35-rados-jewel---basic-smithi/ | ||
rgw | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:08:36-rgw-jewel-distro-basic-smithi/ | PASSED |
http://pulpito.ceph.com/yuriw-2017-02-24_00:03:16-rgw-jewel---basic-smithi/ | ||
rbd | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:19:29-rbd-jewel-distro-basic-smithi/ | PASSED |
http://pulpito.ceph.com/yuriw-2017-02-24_16:59:22-rbd-jewel---basic-smithi/ | ||
fs | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:23:09-fs-jewel-distro-basic-smithi/ | PASSED |
http://pulpito.ceph.com/yuriw-2017-02-24_17:02:22-fs-jewel---basic-smithi/ | ||
kcephfs | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:26:11-kcephfs-jewel-testing-basic-smithi/ | PASSED |
knfs | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:28:17-knfs-jewel-testing-basic-smithi/ | PASSED |
rest | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:29:02-rest-jewel-distro-basic-smithi/ | PASSED |
hadoop | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:29:41-hadoop-jewel-distro-basic-smithi/ | PASSED |
http://pulpito.ceph.com/yuriw-2017-02-24_17:24:15-hadoop-jewel---basic-smithi/ | ||
ceph-deploy | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:30:27-ceph-deploy-jewel-distro-basic-vps/ | PASSED |
ceph-disk | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:31:57-ceph-disk-jewel-distro-basic-vps/ | PASSED |
upgrade/jewel-x/point-to-point-x | http://pulpito.front.sepia.ceph.com:80/yuriw-2017-02-23_21:35:13-upgrade:jewel-x:point-to-point-x-jewel-distro-basic-vps/ | PASSED |
powercycle | http://pulpito.front.sepia.ceph.com/yuriw-2017-02-23_21:04:26-powercycle-jewel-testing-basic-smithi/ | PASSED |
ceph-ansible | http://pulpito.ceph.com/yuriw-2017-02-28_22:57:14-ceph-ansible-jewel-distro-basic-ovh/ | PASSED |
PASSED / FAILED | ||
#93 Updated by Yuri Weinstein over 6 years ago
- Description updated (diff)
#94 Updated by Yuri Weinstein over 6 years ago
- Description updated (diff)
#95 Updated by Nathan Cutler about 6 years ago
- Status changed from In Progress to Resolved
#96 Updated by Nathan Cutler about 6 years ago
- Release set to jewel