Project

General

Profile

ceph.log

Joshua Schmid, 11/09/2016 05:26 PM

Download (12.1 KB)

 
1
salt-master :: ~ » ceph-deploy -v osd prepare --dmcrypt salt-minion-2:/dev/vdb                                                                                                                                  1 ↵
2
 [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
3
[ceph_deploy.cli][INFO  ] Invoked (1.5.34): /usr/bin/ceph-deploy -v osd prepare --dmcrypt salt-minion-2:/dev/vdb
4
[ceph_deploy.cli][INFO  ] ceph-deploy options:
5
[ceph_deploy.cli][INFO  ]  username                      : None
6
[ceph_deploy.cli][INFO  ]  disk                          : [('salt-minion-2', '/dev/vdb', None)]
7
[ceph_deploy.cli][INFO  ]  dmcrypt                       : True
8
[ceph_deploy.cli][INFO  ]  verbose                       : True
9
[ceph_deploy.cli][INFO  ]  bluestore                     : None
10
[ceph_deploy.cli][INFO  ]  overwrite_conf                : False
11
[ceph_deploy.cli][INFO  ]  subcommand                    : prepare
12
[ceph_deploy.cli][INFO  ]  dmcrypt_key_dir               : /etc/ceph/dmcrypt-keys
13
[ceph_deploy.cli][INFO  ]  quiet                         : False
14
[ceph_deploy.cli][INFO  ]  cd_conf                       : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f050c081248>
15
[ceph_deploy.cli][INFO  ]  cluster                       : ceph
16
[ceph_deploy.cli][INFO  ]  fs_type                       : xfs
17
[ceph_deploy.cli][INFO  ]  func                          : <function osd at 0x7f050c4d59b0>
18
[ceph_deploy.cli][INFO  ]  ceph_conf                     : None
19
[ceph_deploy.cli][INFO  ]  default_release               : False
20
[ceph_deploy.cli][INFO  ]  zap_disk                      : False
21
[ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks salt-minion-2:/dev/vdb:
22
[salt-minion-2][DEBUG ] connected to host: salt-minion-2 
23
[salt-minion-2][DEBUG ] detect platform information from remote host
24
[salt-minion-2][DEBUG ] detect machine type
25
[salt-minion-2][DEBUG ] find the location of an executable
26
[ceph_deploy.osd][INFO  ] Distro info: SUSE Linux Enterprise Server 12 x86_64
27
[ceph_deploy.osd][DEBUG ] Deploying osd to salt-minion-2
28
[salt-minion-2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
29
[ceph_deploy.osd][ERROR ] RuntimeError: config file /etc/ceph/ceph.conf exists with different content; use --overwrite-conf to overwrite
30
[ceph_deploy][ERROR ] GenericError: Failed to create 1 OSDs
31

    
32
salt-master :: ~ » ceph-deploy --overwrite-conf -v osd prepare --dmcrypt salt-minion-2:/dev/vdb                                                                                                                 1 ↵
33
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf
34
[ceph_deploy.cli][INFO  ] Invoked (1.5.34): /usr/bin/ceph-deploy --overwrite-conf -v osd prepare --dmcrypt salt-minion-2:/dev/vdb
35
[ceph_deploy.cli][INFO  ] ceph-deploy options:
36
[ceph_deploy.cli][INFO  ]  username                      : None
37
[ceph_deploy.cli][INFO  ]  disk                          : [('salt-minion-2', '/dev/vdb', None)]
38
[ceph_deploy.cli][INFO  ]  dmcrypt                       : True
39
[ceph_deploy.cli][INFO  ]  verbose                       : True
40
[ceph_deploy.cli][INFO  ]  bluestore                     : None
41
[ceph_deploy.cli][INFO  ]  overwrite_conf                : True
42
[ceph_deploy.cli][INFO  ]  subcommand                    : prepare
43
[ceph_deploy.cli][INFO  ]  dmcrypt_key_dir               : /etc/ceph/dmcrypt-keys
44
[ceph_deploy.cli][INFO  ]  quiet                         : False
45
[ceph_deploy.cli][INFO  ]  cd_conf                       : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f1843361248>
46
[ceph_deploy.cli][INFO  ]  cluster                       : ceph
47
[ceph_deploy.cli][INFO  ]  fs_type                       : xfs
48
[ceph_deploy.cli][INFO  ]  func                          : <function osd at 0x7f18437b59b0>
49
[ceph_deploy.cli][INFO  ]  ceph_conf                     : None
50
[ceph_deploy.cli][INFO  ]  default_release               : False
51
[ceph_deploy.cli][INFO  ]  zap_disk                      : False
52
[ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks salt-minion-2:/dev/vdb:
53
[salt-minion-2][DEBUG ] connected to host: salt-minion-2 
54
[salt-minion-2][DEBUG ] detect platform information from remote host
55
[salt-minion-2][DEBUG ] detect machine type
56
[salt-minion-2][DEBUG ] find the location of an executable
57
[ceph_deploy.osd][INFO  ] Distro info: SUSE Linux Enterprise Server 12 x86_64
58
[ceph_deploy.osd][DEBUG ] Deploying osd to salt-minion-2
59
[salt-minion-2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
60
[ceph_deploy.osd][DEBUG ] Preparing host salt-minion-2 disk /dev/vdb journal None activate False
61
[salt-minion-2][DEBUG ] find the location of an executable
62
[salt-minion-2][INFO  ] Running command: /usr/sbin/ceph-disk -v prepare --dmcrypt --dmcrypt-key-dir /etc/ceph/dmcrypt-keys --cluster ceph --fs-type xfs -- /dev/vdb
63
[salt-minion-2][WARNIN] command: Running command: /usr/bin/ceph-osd --cluster=ceph --show-config-value=fsid
64
[salt-minion-2][WARNIN] command: Running command: /usr/bin/ceph-osd --check-allows-journal -i 0 --cluster ceph
65
[salt-minion-2][WARNIN] command: Running command: /usr/bin/ceph-osd --check-wants-journal -i 0 --cluster ceph
66
[salt-minion-2][WARNIN] command: Running command: /usr/bin/ceph-osd --check-needs-journal -i 0 --cluster ceph
67
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
68
[salt-minion-2][WARNIN] set_type: Will colocate journal with data on /dev/vdb
69
[salt-minion-2][WARNIN] command: Running command: /usr/bin/ceph-osd --cluster=ceph --show-config-value=osd_journal_size
70
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
71
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
72
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
73
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
74
[salt-minion-2][WARNIN] set_or_create_partition: Creating osd partition on /dev/vdb
75
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
76
[salt-minion-2][WARNIN] ptype_tobe_for_name: name = lockbox
77
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
78
[salt-minion-2][WARNIN] create_partition: Creating lockbox partition num 3 size 10 on /dev/vdb
79
[salt-minion-2][WARNIN] command_check_call: Running command: /usr/sbin/sgdisk --new=3:0:+10M --change-name=3:ceph lockbox --partition-guid=3:None --typecode=3:fb3aabf9-d25f-47cc-bf5e-721d181642be --mbrtogpt -- /dev/vdb
80
[salt-minion-2][DEBUG ] Creating new GPT entries.
81
[salt-minion-2][DEBUG ] The operation has completed successfully.
82
[salt-minion-2][WARNIN] update_partition: Calling partprobe on created device /dev/vdb
83
[salt-minion-2][WARNIN] command_check_call: Running command: /usr/bin/udevadm settle --timeout=600
84
[salt-minion-2][WARNIN] command: Running command: /usr/bin/flock -s /dev/vdb /usr/sbin/partprobe /dev/vdb
85
[salt-minion-2][WARNIN] command_check_call: Running command: /usr/bin/udevadm settle --timeout=600
86
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
87
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb uuid path is /sys/dev/block/254:16/dm/uuid
88
[salt-minion-2][WARNIN] get_dm_uuid: get_dm_uuid /dev/vdb3 uuid path is /sys/dev/block/254:19/dm/uuid
89
[salt-minion-2][WARNIN] populate: Creating lockbox fs on %s: mkfs -t ext4 /dev/vdb3
90
[salt-minion-2][WARNIN] command_check_call: Running command: /sbin/mkfs -t ext4 /dev/vdb3
91
[salt-minion-2][WARNIN] mke2fs 1.42.11 (09-Jul-2014)
92
[salt-minion-2][DEBUG ] Creating filesystem with 10240 1k blocks and 2560 inodes
93
[salt-minion-2][DEBUG ] Filesystem UUID: d1e214d4-9e52-4dbe-af30-d1684c468f63
94
[salt-minion-2][DEBUG ] Superblock backups stored on blocks: 
95
[salt-minion-2][DEBUG ]         8193
96
[salt-minion-2][DEBUG ] 
97
[salt-minion-2][DEBUG ] Allocating group tables: done                            
98
[salt-minion-2][DEBUG ] Writing inode tables: done                            
99
[salt-minion-2][DEBUG ] Creating journal (1024 blocks): done
100
[salt-minion-2][DEBUG ] Writing superblocks and filesystem accounting information: done
101
[salt-minion-2][DEBUG ] 
102
[salt-minion-2][WARNIN] populate: Mounting lockbox mount -t ext4 /dev/vdb3 /var/lib/ceph/osd-lockbox/ce816935-fb5e-414b-8df8-e0c8973f7fb0
103
[salt-minion-2][WARNIN] command_check_call: Running command: /usr/bin/mount -t ext4 /dev/vdb3 /var/lib/ceph/osd-lockbox/ce816935-fb5e-414b-8df8-e0c8973f7fb0
104
[salt-minion-2][WARNIN] command: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd-lockbox/ce816935-fb5e-414b-8df8-e0c8973f7fb0/osd-uuid.1855.tmp
105
[salt-minion-2][WARNIN] command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_dmcrypt_key_size
106
[salt-minion-2][WARNIN] command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_dmcrypt_type
107
[salt-minion-2][WARNIN] command_check_call: Running command: /usr/bin/ceph config-key put dm-crypt/osd/ce816935-fb5e-414b-8df8-e0c8973f7fb0/luks ZVRXLA0ZU4NsU2oixPPLG9wYsC2jlnztzv3BcXA4jIhSU9L63g/Sjo3RGRfvwHEc3BcowLs2m05MkKds4T0cORw8s8dLYpHnbz3H+Em67oQbj/l6JwSt/ErHJK6fWtVy/TwMxyvfME/ZMy4p/CtKM1t7nlxsUTR+DsojZaSlSpo=
108
[salt-minion-2][WARNIN] 2016-11-09 12:25:56.386822 7f1e6bb85700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin: (2) No such file or directory
109
[salt-minion-2][WARNIN] 2016-11-09 12:25:56.387059 7f1e6bb85700 -1 monclient(hunting): ERROR: missing keyring, cannot use cephx for authentication
110
[salt-minion-2][WARNIN] 2016-11-09 12:25:56.387153 7f1e6bb85700  0 librados: client.admin initialization error (2) No such file or directory
111
[salt-minion-2][WARNIN] Error connecting to cluster: ObjectNotFound
112
[salt-minion-2][WARNIN] Traceback (most recent call last):
113
[salt-minion-2][WARNIN]   File "/usr/sbin/ceph-disk", line 9, in <module>
114
[salt-minion-2][WARNIN]     load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')()
115
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 5009, in run
116
[salt-minion-2][WARNIN]     main(sys.argv[1:])
117
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 4960, in main
118
[salt-minion-2][WARNIN]     args.func(args)
119
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 1793, in main
120
[salt-minion-2][WARNIN]     Prepare.factory(args).prepare()
121
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 1782, in prepare
122
[salt-minion-2][WARNIN]     self.prepare_locked()
123
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 1812, in prepare_locked
124
[salt-minion-2][WARNIN]     self.lockbox.prepare()
125
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 2382, in prepare
126
[salt-minion-2][WARNIN]     self.populate()
127
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 2327, in populate
128
[salt-minion-2][WARNIN]     self.create_key()
129
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 2286, in create_key
130
[salt-minion-2][WARNIN]     base64_key,
131
[salt-minion-2][WARNIN]   File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 440, in command_check_call
132
[salt-minion-2][WARNIN]     return subprocess.check_call(arguments)
133
[salt-minion-2][WARNIN]   File "/usr/lib64/python2.7/subprocess.py", line 540, in check_call
134
[salt-minion-2][WARNIN]     raise CalledProcessError(retcode, cmd)
135
[salt-minion-2][WARNIN] subprocess.CalledProcessError: Command '['/usr/bin/ceph', 'config-key', 'put', 'dm-crypt/osd/ce816935-fb5e-414b-8df8-e0c8973f7fb0/luks', 'ZVRXLA0ZU4NsU2oixPPLG9wYsC2jlnztzv3BcXA4jIhSU9L63g/Sjo3RGRfvwHEc3BcowLs2m05MkKds4T0cORw8s8dLYpHnbz3H+Em67oQbj/l6JwSt/ErHJK6fWtVy/TwMxyvfME/ZMy4p/CtKM1t7nlxsUTR+DsojZaSlSpo=']' returned non-zero exit status 1
136
[salt-minion-2][ERROR ] RuntimeError: command returned non-zero exit status: 1
137
[ceph_deploy.osd][ERROR ] Failed to execute command: /usr/sbin/ceph-disk -v prepare --dmcrypt --dmcrypt-key-dir /etc/ceph/dmcrypt-keys --cluster ceph --fs-type xfs -- /dev/vdb
138
[ceph_deploy][ERROR ] GenericError: Failed to create 1 OSDs
139