diff --git a/ceph.conf b/bak.ceph.conf similarity index 87% rename from ceph.conf rename to bak.ceph.conf index 9a4f3b6..c2d8ad3 100644 --- a/ceph.conf +++ b/bak.ceph.conf @@ -1,9 +1,8 @@ [global] fsid = 29ef4020-303a-4b2e-aa24-a1e20e5ba21c #ms_bind_ipv6 = true -mon_initial_members = pine01, pine02, pine03 -mon_host = [v2:192.168.10.160:3300,v1:192.168.10.160:6789], [v2:192.168.10.161:3300,v1:192.168.10.161:6789], [v2:192.168.10.19:3300,v1:192.168.10.19:6789] - +mon_initial_members = pine01, pine02 +mon_host = [v2:192.168.10.160:3300,v1:192.168.10.160:6789], [v2:192.168.10.161:3300,v1:192.168.10.161:6789] auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx diff --git a/ceph-deploy-ceph.log b/ceph-deploy-ceph.log deleted file mode 100644 index 56b4784..0000000 --- a/ceph-deploy-ceph.log +++ /dev/null @@ -1,5800 +0,0 @@ -[2020-03-03 20:33:08,371][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:33:08,371][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root new pine01 pine02 -[2020-03-03 20:33:08,371][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:33:08,371][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:33:08,371][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] ssh_copykey : True -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] mon : ['pine01', 'pine02'] -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] public_network : None -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] cluster_network : None -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:33:08,372][ceph_deploy.cli][INFO ] fsid : None -[2020-03-03 20:33:08,372][ceph_deploy.new][DEBUG ] Creating new cluster named ceph -[2020-03-03 20:33:08,372][ceph_deploy.new][INFO ] making sure passwordless SSH succeeds -[2020-03-03 20:33:08,386][pine01][DEBUG ] connected to host: lenny -[2020-03-03 20:33:08,390][pine01][INFO ] Running command: ssh -CT -o BatchMode=yes pine01 -[2020-03-03 20:33:09,008][ceph_deploy.new][WARNING] could not connect via SSH -[2020-03-03 20:33:09,009][ceph_deploy.new][INFO ] will connect again with password prompt -[2020-03-03 20:33:10,193][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:33:10,194][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:33:10,339][pine01][DEBUG ] detect machine type -[2020-03-03 20:33:10,366][ceph_deploy.new][INFO ] adding public keys to authorized_keys -[2020-03-03 20:33:10,366][pine01][DEBUG ] append contents to file -[2020-03-03 20:33:12,768][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:33:12,769][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:33:12,912][pine01][DEBUG ] detect machine type -[2020-03-03 20:33:12,933][pine01][DEBUG ] find the location of an executable -[2020-03-03 20:33:12,943][pine01][INFO ] Running command: /bin/ip link show -[2020-03-03 20:33:12,982][pine01][INFO ] Running command: /bin/ip addr show -[2020-03-03 20:33:13,015][pine01][DEBUG ] IP addresses found: [u'192.168.10.160', u'fd87:3937:d2b7:0:ba:c9ff:fe22:3d'] -[2020-03-03 20:33:13,016][ceph_deploy.new][DEBUG ] Resolving host pine01 -[2020-03-03 20:33:13,018][ceph_deploy.new][DEBUG ] Monitor pine01 at 192.168.10.160 -[2020-03-03 20:33:13,018][ceph_deploy.new][INFO ] making sure passwordless SSH succeeds -[2020-03-03 20:33:13,029][pine02][DEBUG ] connected to host: lenny -[2020-03-03 20:33:13,032][pine02][INFO ] Running command: ssh -CT -o BatchMode=yes pine02 -[2020-03-03 20:33:13,247][ceph_deploy.new][WARNING] could not connect via SSH -[2020-03-03 20:33:13,247][ceph_deploy.new][INFO ] will connect again with password prompt -[2020-03-03 20:33:13,919][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:33:13,919][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:33:14,040][pine02][DEBUG ] detect machine type -[2020-03-03 20:33:14,070][ceph_deploy.new][INFO ] adding public keys to authorized_keys -[2020-03-03 20:33:14,070][pine02][DEBUG ] append contents to file -[2020-03-03 20:33:15,594][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:33:15,595][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:33:15,717][pine02][DEBUG ] detect machine type -[2020-03-03 20:33:15,741][pine02][DEBUG ] find the location of an executable -[2020-03-03 20:33:15,754][pine02][INFO ] Running command: /bin/ip link show -[2020-03-03 20:33:15,794][pine02][INFO ] Running command: /bin/ip addr show -[2020-03-03 20:33:15,827][pine02][DEBUG ] IP addresses found: [u'192.168.10.161', u'fd87:3937:d2b7:0:9716:6bcb:617c:e7c6', u'fd87:3937:d2b7::b5f'] -[2020-03-03 20:33:15,827][ceph_deploy.new][DEBUG ] Resolving host pine02 -[2020-03-03 20:33:15,830][ceph_deploy.new][DEBUG ] Monitor pine02 at fd87:3937:d2b7::b5f -[2020-03-03 20:33:15,830][ceph_deploy.new][INFO ] Monitors are IPv6, binding Messenger traffic on IPv6 -[2020-03-03 20:33:15,830][ceph_deploy.new][DEBUG ] Monitor initial members are ['pine01', 'pine02'] -[2020-03-03 20:33:15,830][ceph_deploy.new][DEBUG ] Monitor addrs are ['192.168.10.160', '[fd87:3937:d2b7::b5f]'] -[2020-03-03 20:33:15,830][ceph_deploy.new][DEBUG ] Creating a random mon key... -[2020-03-03 20:33:15,830][ceph_deploy.new][DEBUG ] Writing monitor keyring to ceph.mon.keyring... -[2020-03-03 20:33:15,830][ceph_deploy.new][DEBUG ] Writing initial config to ceph.conf... -[2020-03-03 20:33:22,908][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon create-initial -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] subcommand : create-initial -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] keyrings : None -[2020-03-03 20:33:22,909][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:33:22,910][ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts pine01 pine02 -[2020-03-03 20:33:22,910][ceph_deploy.mon][DEBUG ] detecting platform for host pine01 ... -[2020-03-03 20:33:26,188][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:33:26,188][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:33:26,350][pine01][DEBUG ] detect machine type -[2020-03-03 20:33:26,486][pine01][DEBUG ] find the location of an executable -[2020-03-03 20:33:26,516][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-03 20:33:26,516][pine01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-03 20:33:26,516][pine01][DEBUG ] get remote short hostname -[2020-03-03 20:33:26,573][pine01][DEBUG ] deploying mon to pine01 -[2020-03-03 20:33:26,573][pine01][DEBUG ] get remote short hostname -[2020-03-03 20:33:26,592][pine01][DEBUG ] remote hostname: pine01 -[2020-03-03 20:33:26,643][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:33:26,675][pine01][DEBUG ] create the mon path if it does not exist -[2020-03-03 20:33:26,713][pine01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-pine01/done -[2020-03-03 20:33:26,735][pine01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-pine01/done -[2020-03-03 20:33:26,759][pine01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-pine01.mon.keyring -[2020-03-03 20:33:26,759][pine01][DEBUG ] create the monitor keyring file -[2020-03-03 20:33:26,807][pine01][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i pine01 --keyring /var/lib/ceph/tmp/ceph-pine01.mon.keyring --setuser 64045 --setgroup 64045 -[2020-03-03 20:33:29,481][pine01][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-pine01.mon.keyring -[2020-03-03 20:33:29,518][pine01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-03 20:33:29,545][pine01][DEBUG ] create the init path if it does not exist -[2020-03-03 20:33:29,593][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-03 20:33:36,596][pine01][WARNING] No data was received after 7 seconds, disconnecting... -[2020-03-03 20:33:36,632][pine01][INFO ] Running command: systemctl enable ceph-mon@pine01 -[2020-03-03 20:33:43,634][pine01][WARNING] No data was received after 7 seconds, disconnecting... -[2020-03-03 20:33:43,671][pine01][INFO ] Running command: systemctl start ceph-mon@pine01 -[2020-03-03 20:33:50,674][pine01][WARNING] No data was received after 7 seconds, disconnecting... -[2020-03-03 20:33:52,683][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:33:53,449][pine01][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory -[2020-03-03 20:33:53,449][pine01][WARNING] monitor: mon.pine01, might not be running yet -[2020-03-03 20:33:53,455][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:33:54,221][pine01][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory -[2020-03-03 20:33:54,221][pine01][WARNING] monitor pine01 does not exist in monmap -[2020-03-03 20:33:54,222][pine01][WARNING] neither `public_addr` nor `public_network` keys are defined for monitors -[2020-03-03 20:33:54,222][pine01][WARNING] monitors may not be able to form quorum -[2020-03-03 20:33:54,222][ceph_deploy.mon][DEBUG ] detecting platform for host pine02 ... -[2020-03-03 20:33:55,805][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:33:55,806][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:33:55,948][pine02][DEBUG ] detect machine type -[2020-03-03 20:33:55,975][pine02][DEBUG ] find the location of an executable -[2020-03-03 20:33:55,981][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-03 20:33:55,982][pine02][DEBUG ] determining if provided host has same hostname in remote -[2020-03-03 20:33:55,982][pine02][DEBUG ] get remote short hostname -[2020-03-03 20:33:55,988][pine02][DEBUG ] deploying mon to pine02 -[2020-03-03 20:33:55,988][pine02][DEBUG ] get remote short hostname -[2020-03-03 20:33:55,994][pine02][DEBUG ] remote hostname: pine02 -[2020-03-03 20:33:56,006][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:33:56,018][pine02][DEBUG ] create the mon path if it does not exist -[2020-03-03 20:33:56,024][pine02][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-pine02/done -[2020-03-03 20:33:56,030][pine02][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-pine02/done -[2020-03-03 20:33:56,036][pine02][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-pine02.mon.keyring -[2020-03-03 20:33:56,036][pine02][DEBUG ] create the monitor keyring file -[2020-03-03 20:33:56,052][pine02][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i pine02 --keyring /var/lib/ceph/tmp/ceph-pine02.mon.keyring --setuser 64045 --setgroup 64045 -[2020-03-03 20:33:56,520][pine02][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-pine02.mon.keyring -[2020-03-03 20:33:56,527][pine02][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-03 20:33:56,532][pine02][DEBUG ] create the init path if it does not exist -[2020-03-03 20:33:56,546][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-03 20:33:57,222][pine02][INFO ] Running command: systemctl enable ceph-mon@pine02 -[2020-03-03 20:33:57,898][pine02][INFO ] Running command: systemctl start ceph-mon@pine02 -[2020-03-03 20:33:59,976][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:34:00,792][pine02][DEBUG ] ******************************************************************************** -[2020-03-03 20:34:00,792][pine02][DEBUG ] status for monitor: mon.pine02 -[2020-03-03 20:34:00,792][pine02][DEBUG ] { -[2020-03-03 20:34:00,792][pine02][DEBUG ] "election_epoch": 0, -[2020-03-03 20:34:00,792][pine02][DEBUG ] "extra_probe_peers": [ -[2020-03-03 20:34:00,792][pine02][DEBUG ] { -[2020-03-03 20:34:00,792][pine02][DEBUG ] "addrvec": [ -[2020-03-03 20:34:00,792][pine02][DEBUG ] { -[2020-03-03 20:34:00,793][pine02][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-03 20:34:00,793][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:34:00,793][pine02][DEBUG ] "type": "v2" -[2020-03-03 20:34:00,793][pine02][DEBUG ] }, -[2020-03-03 20:34:00,793][pine02][DEBUG ] { -[2020-03-03 20:34:00,793][pine02][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-03 20:34:00,793][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:34:00,793][pine02][DEBUG ] "type": "v1" -[2020-03-03 20:34:00,793][pine02][DEBUG ] } -[2020-03-03 20:34:00,793][pine02][DEBUG ] ] -[2020-03-03 20:34:00,793][pine02][DEBUG ] } -[2020-03-03 20:34:00,793][pine02][DEBUG ] ], -[2020-03-03 20:34:00,793][pine02][DEBUG ] "feature_map": { -[2020-03-03 20:34:00,793][pine02][DEBUG ] "mon": [ -[2020-03-03 20:34:00,793][pine02][DEBUG ] { -[2020-03-03 20:34:00,793][pine02][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-03 20:34:00,793][pine02][DEBUG ] "num": 1, -[2020-03-03 20:34:00,794][pine02][DEBUG ] "release": "luminous" -[2020-03-03 20:34:00,794][pine02][DEBUG ] } -[2020-03-03 20:34:00,794][pine02][DEBUG ] ] -[2020-03-03 20:34:00,794][pine02][DEBUG ] }, -[2020-03-03 20:34:00,794][pine02][DEBUG ] "features": { -[2020-03-03 20:34:00,794][pine02][DEBUG ] "quorum_con": "0", -[2020-03-03 20:34:00,794][pine02][DEBUG ] "quorum_mon": [], -[2020-03-03 20:34:00,794][pine02][DEBUG ] "required_con": "0", -[2020-03-03 20:34:00,794][pine02][DEBUG ] "required_mon": [] -[2020-03-03 20:34:00,794][pine02][DEBUG ] }, -[2020-03-03 20:34:00,794][pine02][DEBUG ] "monmap": { -[2020-03-03 20:34:00,794][pine02][DEBUG ] "created": "2020-03-03 20:33:56.229076", -[2020-03-03 20:34:00,794][pine02][DEBUG ] "epoch": 0, -[2020-03-03 20:34:00,794][pine02][DEBUG ] "features": { -[2020-03-03 20:34:00,794][pine02][DEBUG ] "optional": [], -[2020-03-03 20:34:00,794][pine02][DEBUG ] "persistent": [] -[2020-03-03 20:34:00,794][pine02][DEBUG ] }, -[2020-03-03 20:34:00,795][pine02][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-03 20:34:00,795][pine02][DEBUG ] "min_mon_release": 0, -[2020-03-03 20:34:00,795][pine02][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-03 20:34:00,795][pine02][DEBUG ] "modified": "2020-03-03 20:33:56.229076", -[2020-03-03 20:34:00,795][pine02][DEBUG ] "mons": [ -[2020-03-03 20:34:00,795][pine02][DEBUG ] { -[2020-03-03 20:34:00,795][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-03 20:34:00,795][pine02][DEBUG ] "name": "pine02", -[2020-03-03 20:34:00,795][pine02][DEBUG ] "public_addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-03 20:34:00,795][pine02][DEBUG ] "public_addrs": { -[2020-03-03 20:34:00,795][pine02][DEBUG ] "addrvec": [ -[2020-03-03 20:34:00,795][pine02][DEBUG ] { -[2020-03-03 20:34:00,795][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:3300", -[2020-03-03 20:34:00,795][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:34:00,795][pine02][DEBUG ] "type": "v2" -[2020-03-03 20:34:00,796][pine02][DEBUG ] }, -[2020-03-03 20:34:00,796][pine02][DEBUG ] { -[2020-03-03 20:34:00,796][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789", -[2020-03-03 20:34:00,796][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:34:00,796][pine02][DEBUG ] "type": "v1" -[2020-03-03 20:34:00,796][pine02][DEBUG ] } -[2020-03-03 20:34:00,796][pine02][DEBUG ] ] -[2020-03-03 20:34:00,796][pine02][DEBUG ] }, -[2020-03-03 20:34:00,796][pine02][DEBUG ] "rank": 0 -[2020-03-03 20:34:00,796][pine02][DEBUG ] }, -[2020-03-03 20:34:00,796][pine02][DEBUG ] { -[2020-03-03 20:34:00,796][pine02][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-03 20:34:00,796][pine02][DEBUG ] "name": "pine01", -[2020-03-03 20:34:00,796][pine02][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-03 20:34:00,796][pine02][DEBUG ] "public_addrs": { -[2020-03-03 20:34:00,796][pine02][DEBUG ] "addrvec": [ -[2020-03-03 20:34:00,797][pine02][DEBUG ] { -[2020-03-03 20:34:00,797][pine02][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-03 20:34:00,797][pine02][DEBUG ] "nonce": 1, -[2020-03-03 20:34:00,797][pine02][DEBUG ] "type": "v1" -[2020-03-03 20:34:00,797][pine02][DEBUG ] } -[2020-03-03 20:34:00,797][pine02][DEBUG ] ] -[2020-03-03 20:34:00,797][pine02][DEBUG ] }, -[2020-03-03 20:34:00,797][pine02][DEBUG ] "rank": 1 -[2020-03-03 20:34:00,797][pine02][DEBUG ] } -[2020-03-03 20:34:00,797][pine02][DEBUG ] ] -[2020-03-03 20:34:00,797][pine02][DEBUG ] }, -[2020-03-03 20:34:00,797][pine02][DEBUG ] "name": "pine02", -[2020-03-03 20:34:00,797][pine02][DEBUG ] "outside_quorum": [ -[2020-03-03 20:34:00,797][pine02][DEBUG ] "pine02" -[2020-03-03 20:34:00,797][pine02][DEBUG ] ], -[2020-03-03 20:34:00,797][pine02][DEBUG ] "quorum": [], -[2020-03-03 20:34:00,798][pine02][DEBUG ] "rank": 0, -[2020-03-03 20:34:00,798][pine02][DEBUG ] "state": "probing", -[2020-03-03 20:34:00,798][pine02][DEBUG ] "sync_provider": [] -[2020-03-03 20:34:00,798][pine02][DEBUG ] } -[2020-03-03 20:34:00,798][pine02][DEBUG ] ******************************************************************************** -[2020-03-03 20:34:00,798][pine02][INFO ] monitor: mon.pine02 is running -[2020-03-03 20:34:00,806][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:34:01,522][ceph_deploy.mon][INFO ] processing monitor mon.pine01 -[2020-03-03 20:34:53,559][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root disk zap ebin02 /dev/sda -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] subcommand : zap -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] host : ebin02 -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:34:53,560][ceph_deploy.cli][INFO ] disk : ['/dev/sda'] -[2020-03-03 20:34:53,561][ceph_deploy.osd][DEBUG ] zapping /dev/sda on ebin02 -[2020-03-03 20:34:53,892][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:34:53,892][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:34:54,036][pine01][DEBUG ] detect machine type -[2020-03-03 20:34:54,056][pine01][DEBUG ] find the location of an executable -[2020-03-03 20:34:54,066][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:34:54,832][pine01][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory -[2020-03-03 20:34:54,832][ceph_deploy.mon][WARNING] mon.pine01 monitor is not yet in quorum, tries left: 5 -[2020-03-03 20:34:54,832][ceph_deploy.mon][WARNING] waiting 5 seconds before retrying -[2020-03-03 20:34:55,858][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 20:34:55,858][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 20:34:56,054][ebin02][DEBUG ] detect machine type -[2020-03-03 20:34:56,083][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:34:56,090][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:34:56,091][ebin02][DEBUG ] zeroing last few blocks of device -[2020-03-03 20:34:56,097][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:34:56,114][ebin02][INFO ] Running command: /usr/sbin/ceph-volume lvm zap /dev/sda -[2020-03-03 20:34:59,842][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:35:00,608][pine01][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory -[2020-03-03 20:35:00,608][ceph_deploy.mon][WARNING] mon.pine01 monitor is not yet in quorum, tries left: 4 -[2020-03-03 20:35:00,608][ceph_deploy.mon][WARNING] waiting 10 seconds before retrying -[2020-03-03 20:35:10,656][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:35:11,472][pine01][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory -[2020-03-03 20:35:11,472][ceph_deploy.mon][WARNING] mon.pine01 monitor is not yet in quorum, tries left: 3 -[2020-03-03 20:35:11,472][ceph_deploy.mon][WARNING] waiting 10 seconds before retrying -[2020-03-03 20:35:21,514][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:35:22,330][pine01][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory -[2020-03-03 20:35:22,330][ceph_deploy.mon][WARNING] mon.pine01 monitor is not yet in quorum, tries left: 2 -[2020-03-03 20:35:22,330][ceph_deploy.mon][WARNING] waiting 15 seconds before retrying -[2020-03-03 20:35:37,350][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:35:38,116][pine01][ERROR ] admin_socket: exception getting command descriptions: [Errno 2] No such file or directory -[2020-03-03 20:35:38,116][ceph_deploy.mon][WARNING] mon.pine01 monitor is not yet in quorum, tries left: 1 -[2020-03-03 20:35:38,116][ceph_deploy.mon][WARNING] waiting 20 seconds before retrying -[2020-03-03 20:35:43,876][ebin02][WARNING] --> Zapping: /dev/sda -[2020-03-03 20:35:43,876][ebin02][WARNING] Running command: /bin/dd if=/dev/zero of=/dev/ceph-e4e20c91-d359-4365-8866-89819da498da/osd-block-aecef4d2-31d5-4e48-bdb8-30b94fc461bb bs=1M count=10 -[2020-03-03 20:35:43,876][ebin02][WARNING] stderr: 10+0 records in -[2020-03-03 20:35:43,876][ebin02][WARNING] 10+0 records out -[2020-03-03 20:35:43,884][ebin02][WARNING] stderr: 10485760 bytes (10 MB, 10 MiB) copied, 0.101732 s, 103 MB/s -[2020-03-03 20:35:43,887][ebin02][WARNING] --> --destroy was not specified, but zapping a whole device will remove the partition table -[2020-03-03 20:35:43,894][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,902][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,905][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,912][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,916][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,923][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,926][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,934][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,941][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,945][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,946][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,949][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,949][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,952][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,953][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:35:43,953][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:35:43,956][ebin02][WARNING] --> RuntimeError: could not complete wipefs on device: /dev/sda -[2020-03-03 20:35:44,120][ebin02][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-03 20:35:44,121][ceph_deploy][ERROR ] RuntimeError: Failed to execute command: /usr/sbin/ceph-volume lvm zap /dev/sda - -[2020-03-03 20:35:58,136][ceph_deploy.mon][INFO ] processing monitor mon.pine02 -[2020-03-03 20:35:59,627][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:35:59,627][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:35:59,749][pine02][DEBUG ] detect machine type -[2020-03-03 20:35:59,773][pine02][DEBUG ] find the location of an executable -[2020-03-03 20:35:59,786][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:36:00,552][ceph_deploy.mon][WARNING] mon.pine02 monitor is not yet in quorum, tries left: 5 -[2020-03-03 20:36:00,552][ceph_deploy.mon][WARNING] waiting 5 seconds before retrying -[2020-03-03 20:36:05,565][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:36:06,381][ceph_deploy.mon][WARNING] mon.pine02 monitor is not yet in quorum, tries left: 4 -[2020-03-03 20:36:06,382][ceph_deploy.mon][WARNING] waiting 10 seconds before retrying -[2020-03-03 20:36:16,397][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:36:17,213][ceph_deploy.mon][WARNING] mon.pine02 monitor is not yet in quorum, tries left: 3 -[2020-03-03 20:36:17,213][ceph_deploy.mon][WARNING] waiting 10 seconds before retrying -[2020-03-03 20:36:22,225][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root disk zap ebin02 /dev/sdb -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] subcommand : zap -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:36:22,225][ceph_deploy.cli][INFO ] host : ebin02 -[2020-03-03 20:36:22,226][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:36:22,226][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:36:22,226][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:36:22,226][ceph_deploy.cli][INFO ] disk : ['/dev/sdb'] -[2020-03-03 20:36:22,226][ceph_deploy.osd][DEBUG ] zapping /dev/sdb on ebin02 -[2020-03-03 20:36:24,630][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 20:36:24,631][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 20:36:24,791][ebin02][DEBUG ] detect machine type -[2020-03-03 20:36:24,821][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:36:24,827][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:36:24,827][ebin02][DEBUG ] zeroing last few blocks of device -[2020-03-03 20:36:24,835][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:36:24,852][ebin02][INFO ] Running command: /usr/sbin/ceph-volume lvm zap /dev/sdb -[2020-03-03 20:36:27,229][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:36:27,994][ceph_deploy.mon][WARNING] mon.pine02 monitor is not yet in quorum, tries left: 2 -[2020-03-03 20:36:27,995][ceph_deploy.mon][WARNING] waiting 15 seconds before retrying -[2020-03-03 20:36:43,057][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:36:43,874][ceph_deploy.mon][WARNING] mon.pine02 monitor is not yet in quorum, tries left: 1 -[2020-03-03 20:36:43,874][ceph_deploy.mon][WARNING] waiting 20 seconds before retrying -[2020-03-03 20:37:03,894][ceph_deploy.mon][ERROR ] Some monitors have still not reached quorum: -[2020-03-03 20:37:03,894][ceph_deploy.mon][ERROR ] pine01 -[2020-03-03 20:37:03,894][ceph_deploy.mon][ERROR ] pine02 -[2020-03-03 20:37:12,161][ebin02][WARNING] --> Zapping: /dev/sdb -[2020-03-03 20:37:12,161][ebin02][WARNING] Running command: /bin/dd if=/dev/zero of=/dev/ceph-23610a45-013b-443a-8b68-689a670ca012/osd-block-7bf796f1-dadd-470b-949b-7c5a61693a89 bs=1M count=10 -[2020-03-03 20:37:12,162][ebin02][WARNING] stderr: 10+0 records in -[2020-03-03 20:37:12,162][ebin02][WARNING] 10+0 records out -[2020-03-03 20:37:12,162][ebin02][WARNING] 10485760 bytes (10 MB, 10 MiB) copied, 0.15268 s, 68.7 MB/s -[2020-03-03 20:37:12,166][ebin02][WARNING] --> --destroy was not specified, but zapping a whole device will remove the partition table -[2020-03-03 20:37:12,167][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,167][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,168][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,168][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,168][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,176][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,176][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,177][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,209][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,209][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,217][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,217][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,221][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,221][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,221][ebin02][WARNING] stderr: wipefs: error: /dev/sdb: probing initialization failed: Device or resource busy -[2020-03-03 20:37:12,222][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:37:12,222][ebin02][WARNING] --> RuntimeError: could not complete wipefs on device: /dev/sdb -[2020-03-03 20:37:12,437][ebin02][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-03 20:37:12,438][ceph_deploy][ERROR ] RuntimeError: Failed to execute command: /usr/sbin/ceph-volume lvm zap /dev/sdb - -[2020-03-03 20:50:56,731][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:50:56,731][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon create-initial -[2020-03-03 20:50:56,731][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] subcommand : create-initial -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] keyrings : None -[2020-03-03 20:50:56,732][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:50:56,733][ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts pine01 pine02 -[2020-03-03 20:50:56,733][ceph_deploy.mon][DEBUG ] detecting platform for host pine01 ... -[2020-03-03 20:50:58,409][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:50:58,411][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:50:58,555][pine01][DEBUG ] detect machine type -[2020-03-03 20:50:58,578][pine01][DEBUG ] find the location of an executable -[2020-03-03 20:50:58,583][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-03 20:50:58,584][pine01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-03 20:50:58,584][pine01][DEBUG ] get remote short hostname -[2020-03-03 20:50:58,589][pine01][DEBUG ] deploying mon to pine01 -[2020-03-03 20:50:58,590][pine01][DEBUG ] get remote short hostname -[2020-03-03 20:50:58,594][pine01][DEBUG ] remote hostname: pine01 -[2020-03-03 20:50:58,603][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:50:58,614][pine01][DEBUG ] create the mon path if it does not exist -[2020-03-03 20:50:58,621][pine01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-pine01/done -[2020-03-03 20:50:58,628][pine01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-03 20:50:58,634][pine01][DEBUG ] create the init path if it does not exist -[2020-03-03 20:50:58,648][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-03 20:50:59,445][pine01][INFO ] Running command: systemctl enable ceph-mon@pine01 -[2020-03-03 20:51:00,228][pine01][INFO ] Running command: systemctl start ceph-mon@pine01 -[2020-03-03 20:51:02,312][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:51:03,182][pine01][DEBUG ] ******************************************************************************** -[2020-03-03 20:51:03,183][pine01][DEBUG ] status for monitor: mon.pine01 -[2020-03-03 20:51:03,184][pine01][DEBUG ] { -[2020-03-03 20:51:03,184][pine01][DEBUG ] "election_epoch": 4, -[2020-03-03 20:51:03,185][pine01][DEBUG ] "extra_probe_peers": [ -[2020-03-03 20:51:03,185][pine01][DEBUG ] { -[2020-03-03 20:51:03,185][pine01][DEBUG ] "addrvec": [ -[2020-03-03 20:51:03,185][pine01][DEBUG ] { -[2020-03-03 20:51:03,185][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:3300", -[2020-03-03 20:51:03,186][pine01][DEBUG ] "nonce": 0, -[2020-03-03 20:51:03,186][pine01][DEBUG ] "type": "v2" -[2020-03-03 20:51:03,186][pine01][DEBUG ] }, -[2020-03-03 20:51:03,186][pine01][DEBUG ] { -[2020-03-03 20:51:03,186][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789", -[2020-03-03 20:51:03,187][pine01][DEBUG ] "nonce": 0, -[2020-03-03 20:51:03,187][pine01][DEBUG ] "type": "v1" -[2020-03-03 20:51:03,187][pine01][DEBUG ] } -[2020-03-03 20:51:03,187][pine01][DEBUG ] ] -[2020-03-03 20:51:03,187][pine01][DEBUG ] } -[2020-03-03 20:51:03,188][pine01][DEBUG ] ], -[2020-03-03 20:51:03,188][pine01][DEBUG ] "feature_map": { -[2020-03-03 20:51:03,188][pine01][DEBUG ] "mon": [ -[2020-03-03 20:51:03,188][pine01][DEBUG ] { -[2020-03-03 20:51:03,188][pine01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-03 20:51:03,189][pine01][DEBUG ] "num": 1, -[2020-03-03 20:51:03,189][pine01][DEBUG ] "release": "luminous" -[2020-03-03 20:51:03,189][pine01][DEBUG ] } -[2020-03-03 20:51:03,189][pine01][DEBUG ] ] -[2020-03-03 20:51:03,189][pine01][DEBUG ] }, -[2020-03-03 20:51:03,190][pine01][DEBUG ] "features": { -[2020-03-03 20:51:03,190][pine01][DEBUG ] "quorum_con": "4611087854031667199", -[2020-03-03 20:51:03,190][pine01][DEBUG ] "quorum_mon": [ -[2020-03-03 20:51:03,190][pine01][DEBUG ] "kraken", -[2020-03-03 20:51:03,190][pine01][DEBUG ] "luminous", -[2020-03-03 20:51:03,191][pine01][DEBUG ] "mimic", -[2020-03-03 20:51:03,191][pine01][DEBUG ] "osdmap-prune", -[2020-03-03 20:51:03,191][pine01][DEBUG ] "nautilus" -[2020-03-03 20:51:03,191][pine01][DEBUG ] ], -[2020-03-03 20:51:03,191][pine01][DEBUG ] "required_con": "2449958747315912708", -[2020-03-03 20:51:03,192][pine01][DEBUG ] "required_mon": [ -[2020-03-03 20:51:03,192][pine01][DEBUG ] "kraken", -[2020-03-03 20:51:03,192][pine01][DEBUG ] "luminous", -[2020-03-03 20:51:03,192][pine01][DEBUG ] "mimic", -[2020-03-03 20:51:03,192][pine01][DEBUG ] "osdmap-prune", -[2020-03-03 20:51:03,193][pine01][DEBUG ] "nautilus" -[2020-03-03 20:51:03,193][pine01][DEBUG ] ] -[2020-03-03 20:51:03,193][pine01][DEBUG ] }, -[2020-03-03 20:51:03,193][pine01][DEBUG ] "monmap": { -[2020-03-03 20:51:03,193][pine01][DEBUG ] "created": "2020-03-03 20:33:27.003562", -[2020-03-03 20:51:03,194][pine01][DEBUG ] "epoch": 1, -[2020-03-03 20:51:03,194][pine01][DEBUG ] "features": { -[2020-03-03 20:51:03,194][pine01][DEBUG ] "optional": [], -[2020-03-03 20:51:03,194][pine01][DEBUG ] "persistent": [ -[2020-03-03 20:51:03,194][pine01][DEBUG ] "kraken", -[2020-03-03 20:51:03,195][pine01][DEBUG ] "luminous", -[2020-03-03 20:51:03,195][pine01][DEBUG ] "mimic", -[2020-03-03 20:51:03,195][pine01][DEBUG ] "osdmap-prune", -[2020-03-03 20:51:03,195][pine01][DEBUG ] "nautilus" -[2020-03-03 20:51:03,195][pine01][DEBUG ] ] -[2020-03-03 20:51:03,196][pine01][DEBUG ] }, -[2020-03-03 20:51:03,196][pine01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-03 20:51:03,196][pine01][DEBUG ] "min_mon_release": 14, -[2020-03-03 20:51:03,196][pine01][DEBUG ] "min_mon_release_name": "nautilus", -[2020-03-03 20:51:03,196][pine01][DEBUG ] "modified": "2020-03-03 20:33:27.003562", -[2020-03-03 20:51:03,197][pine01][DEBUG ] "mons": [ -[2020-03-03 20:51:03,197][pine01][DEBUG ] { -[2020-03-03 20:51:03,197][pine01][DEBUG ] "addr": "192.168.10.160:6789/0", -[2020-03-03 20:51:03,197][pine01][DEBUG ] "name": "pine01", -[2020-03-03 20:51:03,197][pine01][DEBUG ] "public_addr": "192.168.10.160:6789/0", -[2020-03-03 20:51:03,198][pine01][DEBUG ] "public_addrs": { -[2020-03-03 20:51:03,198][pine01][DEBUG ] "addrvec": [ -[2020-03-03 20:51:03,198][pine01][DEBUG ] { -[2020-03-03 20:51:03,198][pine01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-03 20:51:03,198][pine01][DEBUG ] "nonce": 0, -[2020-03-03 20:51:03,198][pine01][DEBUG ] "type": "v2" -[2020-03-03 20:51:03,199][pine01][DEBUG ] }, -[2020-03-03 20:51:03,199][pine01][DEBUG ] { -[2020-03-03 20:51:03,199][pine01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-03 20:51:03,199][pine01][DEBUG ] "nonce": 0, -[2020-03-03 20:51:03,199][pine01][DEBUG ] "type": "v1" -[2020-03-03 20:51:03,200][pine01][DEBUG ] } -[2020-03-03 20:51:03,200][pine01][DEBUG ] ] -[2020-03-03 20:51:03,200][pine01][DEBUG ] }, -[2020-03-03 20:51:03,200][pine01][DEBUG ] "rank": 0 -[2020-03-03 20:51:03,200][pine01][DEBUG ] }, -[2020-03-03 20:51:03,201][pine01][DEBUG ] { -[2020-03-03 20:51:03,201][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-03 20:51:03,201][pine01][DEBUG ] "name": "pine02", -[2020-03-03 20:51:03,201][pine01][DEBUG ] "public_addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-03 20:51:03,201][pine01][DEBUG ] "public_addrs": { -[2020-03-03 20:51:03,202][pine01][DEBUG ] "addrvec": [ -[2020-03-03 20:51:03,202][pine01][DEBUG ] { -[2020-03-03 20:51:03,202][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:3300", -[2020-03-03 20:51:03,202][pine01][DEBUG ] "nonce": 0, -[2020-03-03 20:51:03,202][pine01][DEBUG ] "type": "v2" -[2020-03-03 20:51:03,203][pine01][DEBUG ] }, -[2020-03-03 20:51:03,203][pine01][DEBUG ] { -[2020-03-03 20:51:03,203][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789", -[2020-03-03 20:51:03,203][pine01][DEBUG ] "nonce": 0, -[2020-03-03 20:51:03,203][pine01][DEBUG ] "type": "v1" -[2020-03-03 20:51:03,203][pine01][DEBUG ] } -[2020-03-03 20:51:03,204][pine01][DEBUG ] ] -[2020-03-03 20:51:03,204][pine01][DEBUG ] }, -[2020-03-03 20:51:03,204][pine01][DEBUG ] "rank": 1 -[2020-03-03 20:51:03,204][pine01][DEBUG ] } -[2020-03-03 20:51:03,204][pine01][DEBUG ] ] -[2020-03-03 20:51:03,205][pine01][DEBUG ] }, -[2020-03-03 20:51:03,205][pine01][DEBUG ] "name": "pine01", -[2020-03-03 20:51:03,205][pine01][DEBUG ] "outside_quorum": [], -[2020-03-03 20:51:03,205][pine01][DEBUG ] "quorum": [ -[2020-03-03 20:51:03,205][pine01][DEBUG ] 0, -[2020-03-03 20:51:03,205][pine01][DEBUG ] 1 -[2020-03-03 20:51:03,205][pine01][DEBUG ] ], -[2020-03-03 20:51:03,206][pine01][DEBUG ] "quorum_age": 34, -[2020-03-03 20:51:03,206][pine01][DEBUG ] "rank": 0, -[2020-03-03 20:51:03,206][pine01][DEBUG ] "state": "leader", -[2020-03-03 20:51:03,206][pine01][DEBUG ] "sync_provider": [] -[2020-03-03 20:51:03,206][pine01][DEBUG ] } -[2020-03-03 20:51:03,206][pine01][DEBUG ] ******************************************************************************** -[2020-03-03 20:51:03,206][pine01][INFO ] monitor: mon.pine01 is running -[2020-03-03 20:51:03,214][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:51:04,083][ceph_deploy.mon][DEBUG ] detecting platform for host pine02 ... -[2020-03-03 20:51:05,662][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:51:05,664][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:51:05,785][pine02][DEBUG ] detect machine type -[2020-03-03 20:51:05,812][pine02][DEBUG ] find the location of an executable -[2020-03-03 20:51:05,819][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-03 20:51:05,819][pine02][DEBUG ] determining if provided host has same hostname in remote -[2020-03-03 20:51:05,820][pine02][DEBUG ] get remote short hostname -[2020-03-03 20:51:05,826][pine02][DEBUG ] deploying mon to pine02 -[2020-03-03 20:51:05,827][pine02][DEBUG ] get remote short hostname -[2020-03-03 20:51:05,834][pine02][DEBUG ] remote hostname: pine02 -[2020-03-03 20:51:05,846][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:51:05,858][pine02][DEBUG ] create the mon path if it does not exist -[2020-03-03 20:51:05,867][pine02][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-pine02/done -[2020-03-03 20:51:05,874][pine02][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-03 20:51:05,880][pine02][DEBUG ] create the init path if it does not exist -[2020-03-03 20:51:05,898][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-03 20:51:06,586][pine02][INFO ] Running command: systemctl enable ceph-mon@pine02 -[2020-03-03 20:51:07,269][pine02][INFO ] Running command: systemctl start ceph-mon@pine02 -[2020-03-03 20:51:09,354][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:51:10,123][pine02][DEBUG ] ******************************************************************************** -[2020-03-03 20:51:10,123][pine02][DEBUG ] status for monitor: mon.pine02 -[2020-03-03 20:51:10,123][pine02][DEBUG ] { -[2020-03-03 20:51:10,123][pine02][DEBUG ] "election_epoch": 4, -[2020-03-03 20:51:10,123][pine02][DEBUG ] "extra_probe_peers": [ -[2020-03-03 20:51:10,123][pine02][DEBUG ] { -[2020-03-03 20:51:10,123][pine02][DEBUG ] "addrvec": [ -[2020-03-03 20:51:10,123][pine02][DEBUG ] { -[2020-03-03 20:51:10,124][pine02][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-03 20:51:10,124][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:51:10,124][pine02][DEBUG ] "type": "v2" -[2020-03-03 20:51:10,124][pine02][DEBUG ] }, -[2020-03-03 20:51:10,124][pine02][DEBUG ] { -[2020-03-03 20:51:10,124][pine02][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-03 20:51:10,124][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:51:10,124][pine02][DEBUG ] "type": "v1" -[2020-03-03 20:51:10,124][pine02][DEBUG ] } -[2020-03-03 20:51:10,124][pine02][DEBUG ] ] -[2020-03-03 20:51:10,124][pine02][DEBUG ] } -[2020-03-03 20:51:10,124][pine02][DEBUG ] ], -[2020-03-03 20:51:10,124][pine02][DEBUG ] "feature_map": { -[2020-03-03 20:51:10,124][pine02][DEBUG ] "mon": [ -[2020-03-03 20:51:10,124][pine02][DEBUG ] { -[2020-03-03 20:51:10,124][pine02][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-03 20:51:10,124][pine02][DEBUG ] "num": 1, -[2020-03-03 20:51:10,124][pine02][DEBUG ] "release": "luminous" -[2020-03-03 20:51:10,124][pine02][DEBUG ] } -[2020-03-03 20:51:10,124][pine02][DEBUG ] ] -[2020-03-03 20:51:10,124][pine02][DEBUG ] }, -[2020-03-03 20:51:10,124][pine02][DEBUG ] "features": { -[2020-03-03 20:51:10,125][pine02][DEBUG ] "quorum_con": "4611087854031667199", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "quorum_mon": [ -[2020-03-03 20:51:10,125][pine02][DEBUG ] "kraken", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "luminous", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "mimic", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "osdmap-prune", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "nautilus" -[2020-03-03 20:51:10,125][pine02][DEBUG ] ], -[2020-03-03 20:51:10,125][pine02][DEBUG ] "required_con": "2449958747315912708", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "required_mon": [ -[2020-03-03 20:51:10,125][pine02][DEBUG ] "kraken", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "luminous", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "mimic", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "osdmap-prune", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "nautilus" -[2020-03-03 20:51:10,125][pine02][DEBUG ] ] -[2020-03-03 20:51:10,125][pine02][DEBUG ] }, -[2020-03-03 20:51:10,125][pine02][DEBUG ] "monmap": { -[2020-03-03 20:51:10,125][pine02][DEBUG ] "created": "2020-03-03 20:33:27.003562", -[2020-03-03 20:51:10,125][pine02][DEBUG ] "epoch": 1, -[2020-03-03 20:51:10,125][pine02][DEBUG ] "features": { -[2020-03-03 20:51:10,125][pine02][DEBUG ] "optional": [], -[2020-03-03 20:51:10,126][pine02][DEBUG ] "persistent": [ -[2020-03-03 20:51:10,126][pine02][DEBUG ] "kraken", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "luminous", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "mimic", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "osdmap-prune", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "nautilus" -[2020-03-03 20:51:10,126][pine02][DEBUG ] ] -[2020-03-03 20:51:10,126][pine02][DEBUG ] }, -[2020-03-03 20:51:10,126][pine02][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "min_mon_release": 14, -[2020-03-03 20:51:10,126][pine02][DEBUG ] "min_mon_release_name": "nautilus", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "modified": "2020-03-03 20:33:27.003562", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "mons": [ -[2020-03-03 20:51:10,126][pine02][DEBUG ] { -[2020-03-03 20:51:10,126][pine02][DEBUG ] "addr": "192.168.10.160:6789/0", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "name": "pine01", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "public_addr": "192.168.10.160:6789/0", -[2020-03-03 20:51:10,126][pine02][DEBUG ] "public_addrs": { -[2020-03-03 20:51:10,126][pine02][DEBUG ] "addrvec": [ -[2020-03-03 20:51:10,126][pine02][DEBUG ] { -[2020-03-03 20:51:10,126][pine02][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-03 20:51:10,127][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:51:10,127][pine02][DEBUG ] "type": "v2" -[2020-03-03 20:51:10,127][pine02][DEBUG ] }, -[2020-03-03 20:51:10,127][pine02][DEBUG ] { -[2020-03-03 20:51:10,127][pine02][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-03 20:51:10,127][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:51:10,127][pine02][DEBUG ] "type": "v1" -[2020-03-03 20:51:10,127][pine02][DEBUG ] } -[2020-03-03 20:51:10,127][pine02][DEBUG ] ] -[2020-03-03 20:51:10,127][pine02][DEBUG ] }, -[2020-03-03 20:51:10,127][pine02][DEBUG ] "rank": 0 -[2020-03-03 20:51:10,127][pine02][DEBUG ] }, -[2020-03-03 20:51:10,127][pine02][DEBUG ] { -[2020-03-03 20:51:10,127][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-03 20:51:10,127][pine02][DEBUG ] "name": "pine02", -[2020-03-03 20:51:10,127][pine02][DEBUG ] "public_addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-03 20:51:10,127][pine02][DEBUG ] "public_addrs": { -[2020-03-03 20:51:10,127][pine02][DEBUG ] "addrvec": [ -[2020-03-03 20:51:10,127][pine02][DEBUG ] { -[2020-03-03 20:51:10,127][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:3300", -[2020-03-03 20:51:10,127][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:51:10,127][pine02][DEBUG ] "type": "v2" -[2020-03-03 20:51:10,128][pine02][DEBUG ] }, -[2020-03-03 20:51:10,128][pine02][DEBUG ] { -[2020-03-03 20:51:10,128][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789", -[2020-03-03 20:51:10,128][pine02][DEBUG ] "nonce": 0, -[2020-03-03 20:51:10,128][pine02][DEBUG ] "type": "v1" -[2020-03-03 20:51:10,128][pine02][DEBUG ] } -[2020-03-03 20:51:10,128][pine02][DEBUG ] ] -[2020-03-03 20:51:10,128][pine02][DEBUG ] }, -[2020-03-03 20:51:10,128][pine02][DEBUG ] "rank": 1 -[2020-03-03 20:51:10,128][pine02][DEBUG ] } -[2020-03-03 20:51:10,128][pine02][DEBUG ] ] -[2020-03-03 20:51:10,128][pine02][DEBUG ] }, -[2020-03-03 20:51:10,128][pine02][DEBUG ] "name": "pine02", -[2020-03-03 20:51:10,128][pine02][DEBUG ] "outside_quorum": [], -[2020-03-03 20:51:10,128][pine02][DEBUG ] "quorum": [ -[2020-03-03 20:51:10,128][pine02][DEBUG ] 0, -[2020-03-03 20:51:10,128][pine02][DEBUG ] 1 -[2020-03-03 20:51:10,128][pine02][DEBUG ] ], -[2020-03-03 20:51:10,128][pine02][DEBUG ] "quorum_age": 41, -[2020-03-03 20:51:10,128][pine02][DEBUG ] "rank": 1, -[2020-03-03 20:51:10,128][pine02][DEBUG ] "state": "peon", -[2020-03-03 20:51:10,128][pine02][DEBUG ] "sync_provider": [] -[2020-03-03 20:51:10,129][pine02][DEBUG ] } -[2020-03-03 20:51:10,129][pine02][DEBUG ] ******************************************************************************** -[2020-03-03 20:51:10,129][pine02][INFO ] monitor: mon.pine02 is running -[2020-03-03 20:51:10,137][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:51:10,857][ceph_deploy.mon][INFO ] processing monitor mon.pine01 -[2020-03-03 20:51:12,552][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:51:12,554][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:51:12,698][pine01][DEBUG ] detect machine type -[2020-03-03 20:51:12,722][pine01][DEBUG ] find the location of an executable -[2020-03-03 20:51:12,737][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:51:13,608][ceph_deploy.mon][INFO ] mon.pine01 monitor has reached quorum! -[2020-03-03 20:51:13,608][ceph_deploy.mon][INFO ] processing monitor mon.pine02 -[2020-03-03 20:51:15,209][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:51:15,211][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:51:15,329][pine02][DEBUG ] detect machine type -[2020-03-03 20:51:15,358][pine02][DEBUG ] find the location of an executable -[2020-03-03 20:51:15,374][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-03 20:51:16,145][ceph_deploy.mon][INFO ] mon.pine02 monitor has reached quorum! -[2020-03-03 20:51:16,146][ceph_deploy.mon][INFO ] all initial monitors are running and have formed quorum -[2020-03-03 20:51:16,146][ceph_deploy.mon][INFO ] Running gatherkeys... -[2020-03-03 20:51:16,152][ceph_deploy.gatherkeys][INFO ] Storing keys in temp directory /tmp/tmp525mzS -[2020-03-03 20:51:17,853][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:51:17,855][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:51:17,998][pine01][DEBUG ] detect machine type -[2020-03-03 20:51:18,022][pine01][DEBUG ] get remote short hostname -[2020-03-03 20:51:18,027][pine01][DEBUG ] fetch remote file -[2020-03-03 20:51:18,041][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --admin-daemon=/var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-03 20:51:18,971][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.admin -[2020-03-03 20:51:21,003][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-mds -[2020-03-03 20:51:23,036][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-mgr -[2020-03-03 20:51:25,070][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-osd -[2020-03-03 20:51:27,100][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-rgw -[2020-03-03 20:51:29,125][ceph_deploy.gatherkeys][INFO ] Storing ceph.client.admin.keyring -[2020-03-03 20:51:29,126][ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-mds.keyring -[2020-03-03 20:51:29,126][ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-mgr.keyring -[2020-03-03 20:51:29,127][ceph_deploy.gatherkeys][INFO ] keyring 'ceph.mon.keyring' already exists -[2020-03-03 20:51:29,127][ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-osd.keyring -[2020-03-03 20:51:29,128][ceph_deploy.gatherkeys][INFO ] Storing ceph.bootstrap-rgw.keyring -[2020-03-03 20:51:29,129][ceph_deploy.gatherkeys][INFO ] Destroy temp directory /tmp/tmp525mzS -[2020-03-03 20:52:25,195][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin pine01 pine02 riot01 -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] client : ['pine01', 'pine02', 'riot01'] -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:52:25,195][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:52:25,196][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:52:25,196][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to pine01 -[2020-03-03 20:52:26,822][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:52:26,824][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:52:26,969][pine01][DEBUG ] detect machine type -[2020-03-03 20:52:26,993][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:52:27,012][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to pine02 -[2020-03-03 20:52:28,624][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:52:28,626][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:52:28,747][pine02][DEBUG ] detect machine type -[2020-03-03 20:52:28,775][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:52:28,796][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-03 20:52:30,321][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 20:52:30,323][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 20:52:30,493][riot01][DEBUG ] detect machine type -[2020-03-03 20:52:30,535][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:52:52,470][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mgr create riot01 -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] mgr : [('riot01', 'riot01')] -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:52:52,470][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:52:52,470][ceph_deploy.mgr][DEBUG ] Deploying mgr, cluster ceph hosts riot01:riot01 -[2020-03-03 20:52:53,954][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 20:52:53,955][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 20:52:54,125][riot01][DEBUG ] detect machine type -[2020-03-03 20:52:54,166][ceph_deploy.mgr][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:52:54,166][ceph_deploy.mgr][DEBUG ] remote host will use systemd -[2020-03-03 20:52:54,168][ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to riot01 -[2020-03-03 20:52:54,168][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:52:54,198][riot01][WARNING] mgr keyring does not exist yet, creating one -[2020-03-03 20:52:54,198][riot01][DEBUG ] create a keyring file -[2020-03-03 20:52:54,213][riot01][DEBUG ] create path recursively if it doesn't exist -[2020-03-03 20:52:54,231][riot01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.riot01 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-riot01/keyring -[2020-03-03 20:52:55,113][riot01][ERROR ] Traceback (most recent call last): -[2020-03-03 20:52:55,113][riot01][ERROR ] File "/usr/bin/ceph", line 1266, in -[2020-03-03 20:52:55,113][riot01][ERROR ] retval = main() -[2020-03-03 20:52:55,114][riot01][ERROR ] File "/usr/bin/ceph", line 979, in main -[2020-03-03 20:52:55,114][riot01][ERROR ] conffile=conffile) -[2020-03-03 20:52:55,114][riot01][ERROR ] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-03 20:52:55,114][riot01][ERROR ] raise Exception("timed out") -[2020-03-03 20:52:55,114][riot01][ERROR ] Exception: timed out -[2020-03-03 20:52:55,115][riot01][ERROR ] exit code from command was: 1 -[2020-03-03 20:52:55,115][ceph_deploy.mgr][ERROR ] could not create mgr -[2020-03-03 20:52:55,115][ceph_deploy][ERROR ] GenericError: Failed to create 1 MGRs - -[2020-03-03 20:53:22,331][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mgr create pine01 pine02 -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] mgr : [('pine01', 'pine01'), ('pine02', 'pine02')] -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:53:22,332][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:53:22,332][ceph_deploy.mgr][DEBUG ] Deploying mgr, cluster ceph hosts pine01:pine01 pine02:pine02 -[2020-03-03 20:53:23,942][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 20:53:23,944][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 20:53:24,089][pine01][DEBUG ] detect machine type -[2020-03-03 20:53:24,113][ceph_deploy.mgr][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:53:24,113][ceph_deploy.mgr][DEBUG ] remote host will use systemd -[2020-03-03 20:53:24,115][ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to pine01 -[2020-03-03 20:53:24,115][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:53:24,130][pine01][WARNING] mgr keyring does not exist yet, creating one -[2020-03-03 20:53:24,131][pine01][DEBUG ] create a keyring file -[2020-03-03 20:53:24,145][pine01][DEBUG ] create path recursively if it doesn't exist -[2020-03-03 20:53:24,158][pine01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.pine01 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-pine01/keyring -[2020-03-03 20:53:26,313][pine01][INFO ] Running command: systemctl enable ceph-mgr@pine01 -[2020-03-03 20:53:27,099][pine01][INFO ] Running command: systemctl start ceph-mgr@pine01 -[2020-03-03 20:53:27,180][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-03 20:53:29,616][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 20:53:29,618][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 20:53:29,736][pine02][DEBUG ] detect machine type -[2020-03-03 20:53:29,762][ceph_deploy.mgr][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:53:29,763][ceph_deploy.mgr][DEBUG ] remote host will use systemd -[2020-03-03 20:53:29,763][ceph_deploy.mgr][DEBUG ] deploying mgr bootstrap to pine02 -[2020-03-03 20:53:29,764][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:53:29,783][pine02][WARNING] mgr keyring does not exist yet, creating one -[2020-03-03 20:53:29,783][pine02][DEBUG ] create a keyring file -[2020-03-03 20:53:29,796][pine02][DEBUG ] create path recursively if it doesn't exist -[2020-03-03 20:53:29,812][pine02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mgr --keyring /var/lib/ceph/bootstrap-mgr/ceph.keyring auth get-or-create mgr.pine02 mon allow profile mgr osd allow * mds allow * -o /var/lib/ceph/mgr/ceph-pine02/keyring -[2020-03-03 20:53:31,614][pine02][INFO ] Running command: systemctl enable ceph-mgr@pine02 -[2020-03-03 20:53:32,298][pine02][INFO ] Running command: systemctl start ceph-mgr@pine02 -[2020-03-03 20:53:32,379][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-03 20:53:44,061][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:53:44,061][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd list ebin01 ebin02 -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] subcommand : list -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] host : ['ebin01', 'ebin02'] -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:53:44,062][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:53:46,390][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 20:53:46,391][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 20:53:46,580][ebin01][DEBUG ] detect machine type -[2020-03-03 20:53:46,611][ebin01][DEBUG ] find the location of an executable -[2020-03-03 20:53:46,619][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:53:46,620][ceph_deploy.osd][DEBUG ] Listing disks on ebin01... -[2020-03-03 20:53:46,620][ebin01][DEBUG ] find the location of an executable -[2020-03-03 20:53:46,639][ebin01][INFO ] Running command: /usr/sbin/ceph-volume lvm list -[2020-03-03 20:53:48,881][ebin01][DEBUG ] -[2020-03-03 20:53:48,882][ebin01][DEBUG ] -[2020-03-03 20:53:48,882][ebin01][DEBUG ] ====== osd.0 ======= -[2020-03-03 20:53:48,883][ebin01][DEBUG ] -[2020-03-03 20:53:48,883][ebin01][DEBUG ] [block] /dev/ceph-0ab81ebc-c87d-4f64-9dd7-a434ee256a64/osd-block-59a20cc8-f77c-4851-a06d-45e57a2562d3 -[2020-03-03 20:53:48,883][ebin01][DEBUG ] -[2020-03-03 20:53:48,883][ebin01][DEBUG ] block device /dev/ceph-0ab81ebc-c87d-4f64-9dd7-a434ee256a64/osd-block-59a20cc8-f77c-4851-a06d-45e57a2562d3 -[2020-03-03 20:53:48,884][ebin01][DEBUG ] block uuid KWoO4Y-Qu5f-fmaq-LY5F-CJwm-Sh9J-rYKPfg -[2020-03-03 20:53:48,884][ebin01][DEBUG ] cephx lockbox secret -[2020-03-03 20:53:48,884][ebin01][DEBUG ] cluster fsid 18a73a87-beb2-4997-95dd-2a42fb758b13 -[2020-03-03 20:53:48,884][ebin01][DEBUG ] cluster name ceph -[2020-03-03 20:53:48,885][ebin01][DEBUG ] crush device class None -[2020-03-03 20:53:48,885][ebin01][DEBUG ] encrypted 0 -[2020-03-03 20:53:48,885][ebin01][DEBUG ] osd fsid 59a20cc8-f77c-4851-a06d-45e57a2562d3 -[2020-03-03 20:53:48,885][ebin01][DEBUG ] osd id 0 -[2020-03-03 20:53:48,885][ebin01][DEBUG ] type block -[2020-03-03 20:53:48,886][ebin01][DEBUG ] vdo 0 -[2020-03-03 20:53:48,886][ebin01][DEBUG ] devices /dev/sda -[2020-03-03 20:53:51,323][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 20:53:51,325][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 20:53:51,483][ebin02][DEBUG ] detect machine type -[2020-03-03 20:53:51,515][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:53:51,523][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:53:51,524][ceph_deploy.osd][DEBUG ] Listing disks on ebin02... -[2020-03-03 20:53:51,524][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:53:51,543][ebin02][INFO ] Running command: /usr/sbin/ceph-volume lvm list -[2020-03-03 20:53:53,475][ebin02][WARNING] No valid Ceph devices found -[2020-03-03 20:53:53,641][ebin02][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-03 20:53:53,642][ceph_deploy][ERROR ] RuntimeError: Failed to execute command: /usr/sbin/ceph-volume lvm list - -[2020-03-03 20:54:01,325][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root disk zap ebin02 /dev/sda -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] subcommand : zap -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] host : ebin02 -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:54:01,325][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:54:01,326][ceph_deploy.cli][INFO ] disk : ['/dev/sda'] -[2020-03-03 20:54:01,326][ceph_deploy.osd][DEBUG ] zapping /dev/sda on ebin02 -[2020-03-03 20:54:03,608][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 20:54:03,610][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 20:54:03,769][ebin02][DEBUG ] detect machine type -[2020-03-03 20:54:03,800][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:54:03,808][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:54:03,808][ebin02][DEBUG ] zeroing last few blocks of device -[2020-03-03 20:54:03,815][ebin02][DEBUG ] find the location of an executable -[2020-03-03 20:54:03,836][ebin02][INFO ] Running command: /usr/sbin/ceph-volume lvm zap /dev/sda -[2020-03-03 20:54:47,536][ebin02][WARNING] --> Zapping: /dev/sda -[2020-03-03 20:54:47,537][ebin02][WARNING] Running command: /bin/dd if=/dev/zero of=/dev/ceph-e4e20c91-d359-4365-8866-89819da498da/osd-block-aecef4d2-31d5-4e48-bdb8-30b94fc461bb bs=1M count=10 -[2020-03-03 20:54:47,538][ebin02][WARNING] stderr: 10+0 records in -[2020-03-03 20:54:47,538][ebin02][WARNING] 10+0 records out -[2020-03-03 20:54:47,538][ebin02][WARNING] 10485760 bytes (10 MB, 10 MiB) copied, 0.118145 s, 88.8 MB/s -[2020-03-03 20:54:47,538][ebin02][WARNING] --> --destroy was not specified, but zapping a whole device will remove the partition table -[2020-03-03 20:54:47,542][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,546][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,550][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,557][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,559][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,567][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,571][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,574][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,578][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,582][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,586][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,594][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,595][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,596][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,598][ebin02][WARNING] stderr: wipefs: error: /dev/sda: probing initialization failed: Device or resource busy -[2020-03-03 20:54:47,601][ebin02][WARNING] --> failed to wipefs device, will try again to workaround probable race condition -[2020-03-03 20:54:47,602][ebin02][WARNING] --> RuntimeError: could not complete wipefs on device: /dev/sda -[2020-03-03 20:54:47,767][ebin02][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-03 20:54:47,768][ceph_deploy][ERROR ] RuntimeError: Failed to execute command: /usr/sbin/ceph-volume lvm zap /dev/sda - -[2020-03-03 20:55:52,509][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sda ebin01 -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] journal : None -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] host : ebin01 -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] filestore : None -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] data : /dev/sda -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] block_db : None -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:55:52,510][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 20:55:52,511][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sda -[2020-03-03 20:55:54,747][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 20:55:54,748][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 20:55:54,908][ebin01][DEBUG ] detect machine type -[2020-03-03 20:55:54,941][ebin01][DEBUG ] find the location of an executable -[2020-03-03 20:55:54,949][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:55:54,949][ceph_deploy.osd][DEBUG ] Deploying osd to ebin01 -[2020-03-03 20:55:54,950][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:55:54,970][ebin01][WARNING] osd keyring does not exist yet, creating one -[2020-03-03 20:55:54,971][ebin01][DEBUG ] create a keyring file -[2020-03-03 20:55:54,986][ebin01][DEBUG ] find the location of an executable -[2020-03-03 20:55:55,003][ebin01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-03 20:56:08,938][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 20:56:08,938][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 6f0f2140-9f0f-4de3-b45a-4aed93f84322 -[2020-03-03 20:56:08,941][ebin01][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-d3f8da3f-65ab-4e91-afa5-fd3722fc9874 /dev/sda -[2020-03-03 20:56:08,942][ebin01][WARNING] stdout: Physical volume "/dev/sda" successfully created. -[2020-03-03 20:56:08,945][ebin01][WARNING] stdout: Volume group "ceph-d3f8da3f-65ab-4e91-afa5-fd3722fc9874" successfully created -[2020-03-03 20:56:08,945][ebin01][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-6f0f2140-9f0f-4de3-b45a-4aed93f84322 ceph-d3f8da3f-65ab-4e91-afa5-fd3722fc9874 -[2020-03-03 20:56:08,945][ebin01][WARNING] stdout: Logical volume "osd-block-6f0f2140-9f0f-4de3-b45a-4aed93f84322" created. -[2020-03-03 20:56:08,949][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 20:56:08,949][ebin01][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 -[2020-03-03 20:56:08,950][ebin01][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-03 20:56:08,953][ebin01][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-03 20:56:08,953][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/ceph-d3f8da3f-65ab-4e91-afa5-fd3722fc9874/osd-block-6f0f2140-9f0f-4de3-b45a-4aed93f84322 -[2020-03-03 20:56:08,954][ebin01][WARNING] stderr: /bin/chown: cannot access '/dev/ceph-d3f8da3f-65ab-4e91-afa5-fd3722fc9874/osd-block-6f0f2140-9f0f-4de3-b45a-4aed93f84322': No such file or directory -[2020-03-03 20:56:08,957][ebin01][WARNING] --> Was unable to complete a new OSD, will rollback changes -[2020-03-03 20:56:08,958][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.0 --yes-i-really-mean-it -[2020-03-03 20:56:08,959][ebin01][WARNING] stderr: 2020-03-03 20:56:06.766 7fa33831e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-03 20:56:08,963][ebin01][WARNING] 2020-03-03 20:56:06.766 7fa33831e0 -1 AuthRegistry(0x7f9c0814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-03 20:56:08,963][ebin01][WARNING] stderr: purged osd.0 -[2020-03-03 20:56:08,964][ebin01][WARNING] --> RuntimeError: command returned non-zero exit status: 1 -[2020-03-03 20:56:09,128][ebin01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-03 20:56:09,128][ceph_deploy.osd][ERROR ] Failed to execute command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-03 20:56:09,128][ceph_deploy][ERROR ] GenericError: Failed to create 1 OSDs - -[2020-03-03 20:59:09,657][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sda ebin01 -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] username : root -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] journal : None -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] host : ebin01 -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] filestore : None -[2020-03-03 20:59:09,658][ceph_deploy.cli][INFO ] func : -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] data : /dev/sda -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] block_db : None -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 20:59:09,659][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 20:59:09,659][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sda -[2020-03-03 20:59:30,781][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 20:59:30,783][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 20:59:30,991][ebin01][DEBUG ] detect machine type -[2020-03-03 20:59:31,023][ebin01][DEBUG ] find the location of an executable -[2020-03-03 20:59:31,031][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 20:59:31,032][ceph_deploy.osd][DEBUG ] Deploying osd to ebin01 -[2020-03-03 20:59:31,033][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 20:59:31,459][ebin01][DEBUG ] find the location of an executable -[2020-03-03 20:59:31,481][ebin01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-03 20:59:42,105][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 20:59:42,106][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 72de8b10-759d-4eba-b5b0-db5de0f903e0 -[2020-03-03 20:59:42,106][ebin01][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-cc5335d3-3c90-46b8-9d21-c510e671aa48 /dev/sda -[2020-03-03 20:59:42,106][ebin01][WARNING] stderr: Physical volume '/dev/sda' is already in volume group 'ceph-d3f8da3f-65ab-4e91-afa5-fd3722fc9874' -[2020-03-03 20:59:42,106][ebin01][WARNING] Unable to add physical volume '/dev/sda' to volume group 'ceph-d3f8da3f-65ab-4e91-afa5-fd3722fc9874' -[2020-03-03 20:59:42,106][ebin01][WARNING] /dev/sda: physical volume not initialized. -[2020-03-03 20:59:42,106][ebin01][WARNING] --> Was unable to complete a new OSD, will rollback changes -[2020-03-03 20:59:42,106][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.0 --yes-i-really-mean-it -[2020-03-03 20:59:42,110][ebin01][WARNING] stderr: 2020-03-03 20:59:39.906 7fad7d91e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-03 20:59:42,126][ebin01][WARNING] 2020-03-03 20:59:39.906 7fad7d91e0 -1 AuthRegistry(0x7fa80814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-03 20:59:42,126][ebin01][WARNING] stderr: purged osd.0 -[2020-03-03 20:59:42,126][ebin01][WARNING] --> RuntimeError: command returned non-zero exit status: 5 -[2020-03-03 20:59:42,291][ebin01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-03 20:59:42,291][ceph_deploy.osd][ERROR ] Failed to execute command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-03 20:59:42,292][ceph_deploy][ERROR ] GenericError: Failed to create 1 OSDs - -[2020-03-03 21:01:10,466][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sda ebin01 -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] journal : None -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] host : ebin01 -[2020-03-03 21:01:10,466][ceph_deploy.cli][INFO ] filestore : None -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] data : /dev/sda -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] block_db : None -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:01:10,467][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 21:01:10,467][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sda -[2020-03-03 21:01:12,702][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 21:01:12,704][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 21:01:12,862][ebin01][DEBUG ] detect machine type -[2020-03-03 21:01:12,894][ebin01][DEBUG ] find the location of an executable -[2020-03-03 21:01:12,902][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 21:01:12,902][ceph_deploy.osd][DEBUG ] Deploying osd to ebin01 -[2020-03-03 21:01:12,903][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:01:12,927][ebin01][DEBUG ] find the location of an executable -[2020-03-03 21:01:12,946][ebin01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-03 21:01:43,207][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:01:43,208][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new ff3a69c1-930c-4128-a640-2f85d0f0a860 -[2020-03-03 21:01:43,209][ebin01][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-3da4301b-1b3e-41f3-943d-475852aa6757 /dev/sda -[2020-03-03 21:01:43,209][ebin01][WARNING] stdout: Physical volume "/dev/sda" successfully created. -[2020-03-03 21:01:43,209][ebin01][WARNING] stdout: Volume group "ceph-3da4301b-1b3e-41f3-943d-475852aa6757" successfully created -[2020-03-03 21:01:43,210][ebin01][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-ff3a69c1-930c-4128-a640-2f85d0f0a860 ceph-3da4301b-1b3e-41f3-943d-475852aa6757 -[2020-03-03 21:01:43,211][ebin01][WARNING] stdout: Logical volume "osd-block-ff3a69c1-930c-4128-a640-2f85d0f0a860" created. -[2020-03-03 21:01:43,211][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:01:43,211][ebin01][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 -[2020-03-03 21:01:43,212][ebin01][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-03 21:01:43,212][ebin01][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-03 21:01:43,212][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-3da4301b-1b3e-41f3-943d-475852aa6757/osd-block-ff3a69c1-930c-4128-a640-2f85d0f0a860 -[2020-03-03 21:01:43,213][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-0 -[2020-03-03 21:01:43,213][ebin01][WARNING] Running command: /bin/ln -s /dev/ceph-3da4301b-1b3e-41f3-943d-475852aa6757/osd-block-ff3a69c1-930c-4128-a640-2f85d0f0a860 /var/lib/ceph/osd/ceph-0/block -[2020-03-03 21:01:43,213][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap -[2020-03-03 21:01:43,214][ebin01][WARNING] stderr: 2020-03-03 21:01:30.006 7faab2a1e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-03 21:01:43,215][ebin01][WARNING] 2020-03-03 21:01:30.006 7faab2a1e0 -1 AuthRegistry(0x7fa40814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-03 21:01:43,215][ebin01][WARNING] stderr: got monmap epoch 1 -[2020-03-03 21:01:43,215][ebin01][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-0/keyring --create-keyring --name osd.0 --add-key AQCMt15e8JhxJBAAUg1f+oCXJZTYVqS2tM3g8w== -[2020-03-03 21:01:43,216][ebin01][WARNING] stdout: creating /var/lib/ceph/osd/ceph-0/keyring -[2020-03-03 21:01:43,216][ebin01][WARNING] added entity osd.0 auth(key=AQCMt15e8JhxJBAAUg1f+oCXJZTYVqS2tM3g8w==) -[2020-03-03 21:01:43,216][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring -[2020-03-03 21:01:43,216][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ -[2020-03-03 21:01:43,217][ebin01][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid ff3a69c1-930c-4128-a640-2f85d0f0a860 --setuser ceph --setgroup ceph -[2020-03-03 21:01:43,217][ebin01][WARNING] --> ceph-volume lvm prepare successful for: /dev/sda -[2020-03-03 21:01:43,217][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 -[2020-03-03 21:01:43,218][ebin01][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-3da4301b-1b3e-41f3-943d-475852aa6757/osd-block-ff3a69c1-930c-4128-a640-2f85d0f0a860 --path /var/lib/ceph/osd/ceph-0 --no-mon-config -[2020-03-03 21:01:43,218][ebin01][WARNING] Running command: /bin/ln -snf /dev/ceph-3da4301b-1b3e-41f3-943d-475852aa6757/osd-block-ff3a69c1-930c-4128-a640-2f85d0f0a860 /var/lib/ceph/osd/ceph-0/block -[2020-03-03 21:01:43,219][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-0/block -[2020-03-03 21:01:43,220][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-0 -[2020-03-03 21:01:43,221][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 -[2020-03-03 21:01:43,225][ebin01][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-0-ff3a69c1-930c-4128-a640-2f85d0f0a860 -[2020-03-03 21:01:43,232][ebin01][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-0-ff3a69c1-930c-4128-a640-2f85d0f0a860.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-03 21:01:43,234][ebin01][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@0 -[2020-03-03 21:01:43,238][ebin01][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@0.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-03 21:01:43,254][ebin01][WARNING] Running command: /bin/systemctl start ceph-osd@0 -[2020-03-03 21:01:43,254][ebin01][WARNING] --> ceph-volume lvm activate successful for osd ID: 0 -[2020-03-03 21:01:43,254][ebin01][WARNING] --> ceph-volume lvm create successful for: /dev/sda -[2020-03-03 21:01:48,424][ebin01][INFO ] checking OSD status... -[2020-03-03 21:01:48,424][ebin01][DEBUG ] find the location of an executable -[2020-03-03 21:01:48,443][ebin01][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-03 21:01:49,614][ceph_deploy.osd][DEBUG ] Host ebin01 is now ready for osd use. -[2020-03-03 21:01:56,256][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:01:56,256][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sdb ebin01 -[2020-03-03 21:01:56,256][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] journal : None -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] host : ebin01 -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] filestore : None -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] data : /dev/sdb -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] block_db : None -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:01:56,257][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 21:01:56,258][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdb -[2020-03-03 21:01:58,457][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 21:01:58,458][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 21:01:58,620][ebin01][DEBUG ] detect machine type -[2020-03-03 21:01:58,654][ebin01][DEBUG ] find the location of an executable -[2020-03-03 21:01:58,662][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 21:01:58,662][ceph_deploy.osd][DEBUG ] Deploying osd to ebin01 -[2020-03-03 21:01:58,663][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:01:58,685][ebin01][DEBUG ] find the location of an executable -[2020-03-03 21:01:58,705][ebin01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdb -[2020-03-03 21:02:19,575][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:02:19,576][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 9afbd9a8-8bf7-4357-a6b0-4be056f5cdab -[2020-03-03 21:02:19,576][ebin01][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-df3d495c-1c7a-4717-80b0-8af7e1e6cd5e /dev/sdb -[2020-03-03 21:02:19,576][ebin01][WARNING] stdout: Physical volume "/dev/sdb" successfully created. -[2020-03-03 21:02:19,577][ebin01][WARNING] stdout: Volume group "ceph-df3d495c-1c7a-4717-80b0-8af7e1e6cd5e" successfully created -[2020-03-03 21:02:19,577][ebin01][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab ceph-df3d495c-1c7a-4717-80b0-8af7e1e6cd5e -[2020-03-03 21:02:19,577][ebin01][WARNING] stdout: Logical volume "osd-block-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab" created. -[2020-03-03 21:02:19,578][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:02:19,578][ebin01][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-1 -[2020-03-03 21:02:19,578][ebin01][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-03 21:02:19,579][ebin01][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-03 21:02:19,579][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-df3d495c-1c7a-4717-80b0-8af7e1e6cd5e/osd-block-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab -[2020-03-03 21:02:19,579][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-1 -[2020-03-03 21:02:19,579][ebin01][WARNING] Running command: /bin/ln -s /dev/ceph-df3d495c-1c7a-4717-80b0-8af7e1e6cd5e/osd-block-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab /var/lib/ceph/osd/ceph-1/block -[2020-03-03 21:02:19,579][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-1/activate.monmap -[2020-03-03 21:02:19,580][ebin01][WARNING] stderr: 2020-03-03 21:02:10.675 7f80e691e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-03 21:02:19,580][ebin01][WARNING] 2020-03-03 21:02:10.675 7f80e691e0 -1 AuthRegistry(0x7f7c0814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-03 21:02:19,580][ebin01][WARNING] stderr: got monmap epoch 1 -[2020-03-03 21:02:19,581][ebin01][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-1/keyring --create-keyring --name osd.1 --add-key AQC5t15ez86/DhAA0HFc9ug+CBI3Su1TcgHLNg== -[2020-03-03 21:02:19,581][ebin01][WARNING] stdout: creating /var/lib/ceph/osd/ceph-1/keyring -[2020-03-03 21:02:19,581][ebin01][WARNING] added entity osd.1 auth(key=AQC5t15ez86/DhAA0HFc9ug+CBI3Su1TcgHLNg==) -[2020-03-03 21:02:19,581][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1/keyring -[2020-03-03 21:02:19,581][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1/ -[2020-03-03 21:02:19,582][ebin01][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 1 --monmap /var/lib/ceph/osd/ceph-1/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-1/ --osd-uuid 9afbd9a8-8bf7-4357-a6b0-4be056f5cdab --setuser ceph --setgroup ceph -[2020-03-03 21:02:19,585][ebin01][WARNING] --> ceph-volume lvm prepare successful for: /dev/sdb -[2020-03-03 21:02:19,585][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 -[2020-03-03 21:02:19,587][ebin01][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-df3d495c-1c7a-4717-80b0-8af7e1e6cd5e/osd-block-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab --path /var/lib/ceph/osd/ceph-1 --no-mon-config -[2020-03-03 21:02:19,591][ebin01][WARNING] Running command: /bin/ln -snf /dev/ceph-df3d495c-1c7a-4717-80b0-8af7e1e6cd5e/osd-block-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab /var/lib/ceph/osd/ceph-1/block -[2020-03-03 21:02:19,593][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-1/block -[2020-03-03 21:02:19,593][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-1 -[2020-03-03 21:02:19,593][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 -[2020-03-03 21:02:19,597][ebin01][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-1-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab -[2020-03-03 21:02:19,597][ebin01][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-1-9afbd9a8-8bf7-4357-a6b0-4be056f5cdab.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-03 21:02:19,605][ebin01][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@1 -[2020-03-03 21:02:19,621][ebin01][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@1.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-03 21:02:19,622][ebin01][WARNING] Running command: /bin/systemctl start ceph-osd@1 -[2020-03-03 21:02:19,622][ebin01][WARNING] --> ceph-volume lvm activate successful for osd ID: 1 -[2020-03-03 21:02:19,622][ebin01][WARNING] --> ceph-volume lvm create successful for: /dev/sdb -[2020-03-03 21:02:24,792][ebin01][INFO ] checking OSD status... -[2020-03-03 21:02:24,792][ebin01][DEBUG ] find the location of an executable -[2020-03-03 21:02:24,807][ebin01][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-03 21:02:25,924][ceph_deploy.osd][DEBUG ] Host ebin01 is now ready for osd use. -[2020-03-03 21:02:31,005][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:02:31,005][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sda ebin02 -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] journal : None -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] host : ebin02 -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] filestore : None -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] data : /dev/sda -[2020-03-03 21:02:31,006][ceph_deploy.cli][INFO ] block_db : None -[2020-03-03 21:02:31,007][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-03 21:02:31,007][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:02:31,007][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-03 21:02:31,007][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:02:31,007][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 21:02:31,007][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sda -[2020-03-03 21:02:33,374][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 21:02:33,374][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 21:02:33,587][ebin02][DEBUG ] detect machine type -[2020-03-03 21:02:33,618][ebin02][DEBUG ] find the location of an executable -[2020-03-03 21:02:33,624][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 21:02:33,625][ceph_deploy.osd][DEBUG ] Deploying osd to ebin02 -[2020-03-03 21:02:33,625][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:02:33,652][ebin02][WARNING] osd keyring does not exist yet, creating one -[2020-03-03 21:02:33,652][ebin02][DEBUG ] create a keyring file -[2020-03-03 21:02:33,669][ebin02][DEBUG ] find the location of an executable -[2020-03-03 21:02:33,686][ebin02][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-03 21:03:02,359][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:03:02,359][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 61771fa8-6d0f-48ec-a03b-6bf891b8eac3 -[2020-03-03 21:03:02,359][ebin02][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-4be08c0e-6cb3-4a39-a6fe-7de4eab46f44 /dev/sda -[2020-03-03 21:03:02,359][ebin02][WARNING] stdout: Physical volume "/dev/sda" successfully created. -[2020-03-03 21:03:02,359][ebin02][WARNING] stdout: Volume group "ceph-4be08c0e-6cb3-4a39-a6fe-7de4eab46f44" successfully created -[2020-03-03 21:03:02,423][ebin02][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-61771fa8-6d0f-48ec-a03b-6bf891b8eac3 ceph-4be08c0e-6cb3-4a39-a6fe-7de4eab46f44 -[2020-03-03 21:03:02,423][ebin02][WARNING] stdout: Logical volume "osd-block-61771fa8-6d0f-48ec-a03b-6bf891b8eac3" created. -[2020-03-03 21:03:02,423][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:03:02,423][ebin02][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-2 -[2020-03-03 21:03:02,423][ebin02][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-03 21:03:02,423][ebin02][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-03 21:03:02,423][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-4be08c0e-6cb3-4a39-a6fe-7de4eab46f44/osd-block-61771fa8-6d0f-48ec-a03b-6bf891b8eac3 -[2020-03-03 21:03:02,423][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-0 -[2020-03-03 21:03:02,423][ebin02][WARNING] Running command: /bin/ln -s /dev/ceph-4be08c0e-6cb3-4a39-a6fe-7de4eab46f44/osd-block-61771fa8-6d0f-48ec-a03b-6bf891b8eac3 /var/lib/ceph/osd/ceph-2/block -[2020-03-03 21:03:02,423][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-2/activate.monmap -[2020-03-03 21:03:02,423][ebin02][WARNING] stderr: 2020-03-03 21:02:45.230 7fa24141e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-03 21:03:02,423][ebin02][WARNING] 2020-03-03 21:02:45.230 7fa24141e0 -1 AuthRegistry(0x7f9c0814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-03 21:03:02,423][ebin02][WARNING] stderr: got monmap epoch 1 -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-2/keyring --create-keyring --name osd.2 --add-key AQDct15eaGUICxAAFnRSbfeV7uCY91xN8+qhbA== -[2020-03-03 21:03:02,424][ebin02][WARNING] stdout: creating /var/lib/ceph/osd/ceph-2/keyring -[2020-03-03 21:03:02,424][ebin02][WARNING] added entity osd.2 auth(key=AQDct15eaGUICxAAFnRSbfeV7uCY91xN8+qhbA==) -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2/keyring -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2/ -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 2 --monmap /var/lib/ceph/osd/ceph-2/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-2/ --osd-uuid 61771fa8-6d0f-48ec-a03b-6bf891b8eac3 --setuser ceph --setgroup ceph -[2020-03-03 21:03:02,424][ebin02][WARNING] --> ceph-volume lvm prepare successful for: /dev/sda -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-4be08c0e-6cb3-4a39-a6fe-7de4eab46f44/osd-block-61771fa8-6d0f-48ec-a03b-6bf891b8eac3 --path /var/lib/ceph/osd/ceph-2 --no-mon-config -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/ln -snf /dev/ceph-4be08c0e-6cb3-4a39-a6fe-7de4eab46f44/osd-block-61771fa8-6d0f-48ec-a03b-6bf891b8eac3 /var/lib/ceph/osd/ceph-2/block -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-2/block -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-0 -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-2-61771fa8-6d0f-48ec-a03b-6bf891b8eac3 -[2020-03-03 21:03:02,424][ebin02][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-2-61771fa8-6d0f-48ec-a03b-6bf891b8eac3.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-03 21:03:02,424][ebin02][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@2 -[2020-03-03 21:03:02,424][ebin02][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@2.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-03 21:03:02,425][ebin02][WARNING] Running command: /bin/systemctl start ceph-osd@2 -[2020-03-03 21:03:02,425][ebin02][WARNING] --> ceph-volume lvm activate successful for osd ID: 2 -[2020-03-03 21:03:02,425][ebin02][WARNING] --> ceph-volume lvm create successful for: /dev/sda -[2020-03-03 21:03:07,428][ebin02][INFO ] checking OSD status... -[2020-03-03 21:03:07,428][ebin02][DEBUG ] find the location of an executable -[2020-03-03 21:03:07,476][ebin02][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-03 21:03:08,492][ceph_deploy.osd][DEBUG ] Host ebin02 is now ready for osd use. -[2020-03-03 21:03:12,755][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sdb ebin02 -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] journal : None -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] host : ebin02 -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] filestore : None -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-03 21:03:12,756][ceph_deploy.cli][INFO ] data : /dev/sdb -[2020-03-03 21:03:12,757][ceph_deploy.cli][INFO ] block_db : None -[2020-03-03 21:03:12,757][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-03 21:03:12,757][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:03:12,757][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-03 21:03:12,757][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:03:12,757][ceph_deploy.cli][INFO ] debug : False -[2020-03-03 21:03:12,757][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdb -[2020-03-03 21:03:17,258][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 21:03:17,259][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 21:03:17,421][ebin02][DEBUG ] detect machine type -[2020-03-03 21:03:17,449][ebin02][DEBUG ] find the location of an executable -[2020-03-03 21:03:17,456][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-03 21:03:17,456][ceph_deploy.osd][DEBUG ] Deploying osd to ebin02 -[2020-03-03 21:03:17,456][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:03:17,476][ebin02][DEBUG ] find the location of an executable -[2020-03-03 21:03:17,494][ebin02][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdb -[2020-03-03 21:03:39,202][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:03:39,202][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 45215366-fe82-4438-ae61-0644cfa457c3 -[2020-03-03 21:03:39,202][ebin02][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-0e6f6cfe-1fe2-46e2-9adf-982e50384074 /dev/sdb -[2020-03-03 21:03:39,202][ebin02][WARNING] stdout: Physical volume "/dev/sdb" successfully created. -[2020-03-03 21:03:39,202][ebin02][WARNING] stdout: Volume group "ceph-0e6f6cfe-1fe2-46e2-9adf-982e50384074" successfully created -[2020-03-03 21:03:39,203][ebin02][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-45215366-fe82-4438-ae61-0644cfa457c3 ceph-0e6f6cfe-1fe2-46e2-9adf-982e50384074 -[2020-03-03 21:03:39,203][ebin02][WARNING] stdout: Logical volume "osd-block-45215366-fe82-4438-ae61-0644cfa457c3" created. -[2020-03-03 21:03:39,203][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-03 21:03:39,203][ebin02][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-3 -[2020-03-03 21:03:39,203][ebin02][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-03 21:03:39,210][ebin02][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-03 21:03:39,210][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-0e6f6cfe-1fe2-46e2-9adf-982e50384074/osd-block-45215366-fe82-4438-ae61-0644cfa457c3 -[2020-03-03 21:03:39,213][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-1 -[2020-03-03 21:03:39,214][ebin02][WARNING] Running command: /bin/ln -s /dev/ceph-0e6f6cfe-1fe2-46e2-9adf-982e50384074/osd-block-45215366-fe82-4438-ae61-0644cfa457c3 /var/lib/ceph/osd/ceph-3/block -[2020-03-03 21:03:39,217][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-3/activate.monmap -[2020-03-03 21:03:39,232][ebin02][WARNING] stderr: 2020-03-03 21:03:30.031 7fb233a1e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-03 21:03:39,240][ebin02][WARNING] 2020-03-03 21:03:30.031 7fb233a1e0 -1 AuthRegistry(0x7fac0814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-03 21:03:39,240][ebin02][WARNING] stderr: got monmap epoch 1 -[2020-03-03 21:03:39,241][ebin02][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-3/keyring --create-keyring --name osd.3 --add-key AQAHuF5eltB9OxAAgUzbsJRUPnWIZWS+yMHIaQ== -[2020-03-03 21:03:39,241][ebin02][WARNING] stdout: creating /var/lib/ceph/osd/ceph-3/keyring -[2020-03-03 21:03:39,244][ebin02][WARNING] added entity osd.3 auth(key=AQAHuF5eltB9OxAAgUzbsJRUPnWIZWS+yMHIaQ==) -[2020-03-03 21:03:39,244][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/keyring -[2020-03-03 21:03:39,248][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3/ -[2020-03-03 21:03:39,249][ebin02][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 3 --monmap /var/lib/ceph/osd/ceph-3/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-3/ --osd-uuid 45215366-fe82-4438-ae61-0644cfa457c3 --setuser ceph --setgroup ceph -[2020-03-03 21:03:39,252][ebin02][WARNING] --> ceph-volume lvm prepare successful for: /dev/sdb -[2020-03-03 21:03:39,255][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 -[2020-03-03 21:03:39,255][ebin02][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-0e6f6cfe-1fe2-46e2-9adf-982e50384074/osd-block-45215366-fe82-4438-ae61-0644cfa457c3 --path /var/lib/ceph/osd/ceph-3 --no-mon-config -[2020-03-03 21:03:39,263][ebin02][WARNING] Running command: /bin/ln -snf /dev/ceph-0e6f6cfe-1fe2-46e2-9adf-982e50384074/osd-block-45215366-fe82-4438-ae61-0644cfa457c3 /var/lib/ceph/osd/ceph-3/block -[2020-03-03 21:03:39,264][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-3/block -[2020-03-03 21:03:39,264][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-1 -[2020-03-03 21:03:39,264][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 -[2020-03-03 21:03:39,272][ebin02][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-3-45215366-fe82-4438-ae61-0644cfa457c3 -[2020-03-03 21:03:39,272][ebin02][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-3-45215366-fe82-4438-ae61-0644cfa457c3.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-03 21:03:39,273][ebin02][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@3 -[2020-03-03 21:03:39,276][ebin02][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@3.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-03 21:03:39,279][ebin02][WARNING] Running command: /bin/systemctl start ceph-osd@3 -[2020-03-03 21:03:39,279][ebin02][WARNING] --> ceph-volume lvm activate successful for osd ID: 3 -[2020-03-03 21:03:39,283][ebin02][WARNING] --> ceph-volume lvm create successful for: /dev/sdb -[2020-03-03 21:03:44,448][ebin02][INFO ] checking OSD status... -[2020-03-03 21:03:44,448][ebin02][DEBUG ] find the location of an executable -[2020-03-03 21:03:45,369][ebin02][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-03 21:03:47,037][ceph_deploy.osd][DEBUG ] Host ebin02 is now ready for osd use. -[2020-03-03 21:09:46,171][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root config push riot01 pine01 pine02 ebin01 ebin02 -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02'] -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:09:46,172][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:09:46,172][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-03 21:09:47,710][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:09:47,712][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:09:47,881][riot01][DEBUG ] detect machine type -[2020-03-03 21:09:47,919][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:09:47,938][ceph_deploy.config][ERROR ] RuntimeError: config file /etc/ceph/ceph.conf exists with different content; use --overwrite-conf to overwrite -[2020-03-03 21:09:47,939][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-03 21:09:49,664][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 21:09:49,666][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 21:09:49,819][pine01][DEBUG ] detect machine type -[2020-03-03 21:09:49,843][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:09:49,855][ceph_deploy.config][ERROR ] RuntimeError: config file /etc/ceph/ceph.conf exists with different content; use --overwrite-conf to overwrite -[2020-03-03 21:09:49,856][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-03 21:09:51,466][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 21:09:51,468][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 21:09:51,588][pine02][DEBUG ] detect machine type -[2020-03-03 21:09:51,612][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:09:51,624][ceph_deploy.config][ERROR ] RuntimeError: config file /etc/ceph/ceph.conf exists with different content; use --overwrite-conf to overwrite -[2020-03-03 21:09:51,624][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-03 21:09:53,661][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 21:09:53,663][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 21:09:53,821][ebin01][DEBUG ] detect machine type -[2020-03-03 21:09:53,854][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:09:53,869][ceph_deploy.config][ERROR ] RuntimeError: config file /etc/ceph/ceph.conf exists with different content; use --overwrite-conf to overwrite -[2020-03-03 21:09:53,870][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-03 21:09:56,045][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 21:09:56,047][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 21:09:56,207][ebin02][DEBUG ] detect machine type -[2020-03-03 21:09:56,239][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:09:56,260][ceph_deploy.config][ERROR ] RuntimeError: config file /etc/ceph/ceph.conf exists with different content; use --overwrite-conf to overwrite -[2020-03-03 21:09:56,261][ceph_deploy][ERROR ] GenericError: Failed to config 5 hosts - -[2020-03-03 21:10:05,807][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:10:05,807][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02'] -[2020-03-03 21:10:05,808][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:10:05,808][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:10:05,808][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:10:05,808][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-03 21:10:07,333][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:10:07,335][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:10:07,504][riot01][DEBUG ] detect machine type -[2020-03-03 21:10:07,545][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:10:07,567][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-03 21:10:09,292][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 21:10:09,294][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 21:10:09,438][pine01][DEBUG ] detect machine type -[2020-03-03 21:10:09,462][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:10:09,473][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-03 21:10:11,090][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 21:10:11,092][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 21:10:11,213][pine02][DEBUG ] detect machine type -[2020-03-03 21:10:11,239][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:10:11,251][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-03 21:10:13,468][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 21:10:13,470][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 21:10:13,632][ebin01][DEBUG ] detect machine type -[2020-03-03 21:10:13,663][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:10:13,679][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-03 21:10:15,896][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 21:10:15,898][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 21:10:16,060][ebin02][DEBUG ] detect machine type -[2020-03-03 21:10:16,092][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:13:47,640][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02'] -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:13:47,641][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:13:47,641][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-03 21:13:49,162][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:13:49,164][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:13:49,337][riot01][DEBUG ] detect machine type -[2020-03-03 21:13:49,382][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:13:49,403][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-03 21:13:51,105][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 21:13:51,107][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 21:13:51,252][pine01][DEBUG ] detect machine type -[2020-03-03 21:13:51,277][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:13:51,289][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-03 21:13:52,910][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 21:13:52,912][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 21:13:53,045][pine02][DEBUG ] detect machine type -[2020-03-03 21:13:53,075][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:13:53,087][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-03 21:13:55,242][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 21:13:55,244][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 21:13:55,403][ebin01][DEBUG ] detect machine type -[2020-03-03 21:13:55,436][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:13:55,451][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-03 21:13:57,594][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 21:13:57,596][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 21:13:57,756][ebin02][DEBUG ] detect machine type -[2020-03-03 21:13:57,790][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:20:14,746][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:20:14,746][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02'] -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:20:14,747][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:20:14,747][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-03 21:20:16,271][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:20:16,273][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:20:16,442][riot01][DEBUG ] detect machine type -[2020-03-03 21:20:16,482][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:20:16,503][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-03 21:20:18,237][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 21:20:18,239][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 21:20:18,395][pine01][DEBUG ] detect machine type -[2020-03-03 21:20:18,424][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:20:18,436][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-03 21:20:20,597][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 21:20:20,597][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 21:20:20,760][pine02][DEBUG ] detect machine type -[2020-03-03 21:20:20,787][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:20:20,800][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-03 21:20:23,224][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 21:20:23,225][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 21:20:23,539][ebin01][DEBUG ] detect machine type -[2020-03-03 21:20:23,585][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:20:23,988][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-03 21:20:27,984][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 21:20:27,986][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 21:20:28,333][ebin02][DEBUG ] detect machine type -[2020-03-03 21:20:28,379][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:27:12,191][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02'] -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:27:12,192][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:27:12,192][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-03 21:27:13,700][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:27:13,702][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:27:13,873][riot01][DEBUG ] detect machine type -[2020-03-03 21:27:13,914][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:27:13,936][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-03 21:27:15,625][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 21:27:15,627][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 21:27:15,773][pine01][DEBUG ] detect machine type -[2020-03-03 21:27:15,797][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:27:15,809][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-03 21:27:17,363][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 21:27:17,365][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 21:27:17,485][pine02][DEBUG ] detect machine type -[2020-03-03 21:27:17,511][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:27:17,525][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-03 21:27:19,746][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 21:27:19,748][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 21:27:19,908][ebin01][DEBUG ] detect machine type -[2020-03-03 21:27:19,940][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:27:19,955][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-03 21:27:22,456][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 21:27:22,458][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 21:27:22,691][ebin02][DEBUG ] detect machine type -[2020-03-03 21:27:22,732][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:42:10,621][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create riot01 -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] rgw : [('riot01', 'rgw.riot01')] -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:42:10,622][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:42:10,622][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts riot01:rgw.riot01 -[2020-03-03 21:42:12,097][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:42:12,099][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:42:12,274][riot01][DEBUG ] detect machine type -[2020-03-03 21:42:12,313][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-03 21:42:12,314][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-03 21:42:12,316][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to riot01 -[2020-03-03 21:42:12,316][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:42:12,343][riot01][WARNING] rgw keyring does not exist yet, creating one -[2020-03-03 21:42:12,343][riot01][DEBUG ] create a keyring file -[2020-03-03 21:42:12,361][riot01][DEBUG ] create path recursively if it doesn't exist -[2020-03-03 21:42:12,378][riot01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.riot01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.riot01/keyring -[2020-03-03 21:42:13,262][riot01][ERROR ] Traceback (most recent call last): -[2020-03-03 21:42:13,262][riot01][ERROR ] File "/usr/bin/ceph", line 1266, in -[2020-03-03 21:42:13,263][riot01][ERROR ] retval = main() -[2020-03-03 21:42:13,263][riot01][ERROR ] File "/usr/bin/ceph", line 979, in main -[2020-03-03 21:42:13,263][riot01][ERROR ] conffile=conffile) -[2020-03-03 21:42:13,263][riot01][ERROR ] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-03 21:42:13,264][riot01][ERROR ] raise Exception("timed out") -[2020-03-03 21:42:13,264][riot01][ERROR ] Exception: timed out -[2020-03-03 21:42:13,264][riot01][ERROR ] exit code from command was: 1 -[2020-03-03 21:42:13,264][ceph_deploy.rgw][ERROR ] could not create rgw -[2020-03-03 21:42:13,265][ceph_deploy][ERROR ] GenericError: Failed to create 1 RGWs - -[2020-03-03 21:43:55,513][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create riot01 -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] rgw : [('riot01', 'rgw.riot01')] -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:43:55,513][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:43:55,514][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts riot01:rgw.riot01 -[2020-03-03 21:43:57,024][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:43:57,026][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:43:57,196][riot01][DEBUG ] detect machine type -[2020-03-03 21:43:57,236][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-03 21:43:57,236][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-03 21:43:57,238][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to riot01 -[2020-03-03 21:43:57,238][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:43:57,268][riot01][DEBUG ] create path recursively if it doesn't exist -[2020-03-03 21:43:57,284][riot01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.riot01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.riot01/keyring -[2020-03-03 21:43:58,167][riot01][ERROR ] Traceback (most recent call last): -[2020-03-03 21:43:58,168][riot01][ERROR ] File "/usr/bin/ceph", line 1266, in -[2020-03-03 21:43:58,168][riot01][ERROR ] retval = main() -[2020-03-03 21:43:58,168][riot01][ERROR ] File "/usr/bin/ceph", line 979, in main -[2020-03-03 21:43:58,168][riot01][ERROR ] conffile=conffile) -[2020-03-03 21:43:58,168][riot01][ERROR ] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-03 21:43:58,169][riot01][ERROR ] raise Exception("timed out") -[2020-03-03 21:43:58,169][riot01][ERROR ] Exception: timed out -[2020-03-03 21:43:58,169][riot01][ERROR ] exit code from command was: 1 -[2020-03-03 21:43:58,169][ceph_deploy.rgw][ERROR ] could not create rgw -[2020-03-03 21:43:58,170][ceph_deploy][ERROR ] GenericError: Failed to create 1 RGWs - -[2020-03-03 21:49:04,635][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin riot01 -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] client : ['riot01'] -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:49:04,635][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:49:04,636][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:49:04,636][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-03 21:49:06,134][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:49:06,137][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:49:06,305][riot01][DEBUG ] detect machine type -[2020-03-03 21:49:06,342][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:49:14,206][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create riot01 -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] username : root -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] rgw : [('riot01', 'rgw.riot01')] -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-03 21:49:14,206][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 21:49:14,207][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 21:49:14,207][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 21:49:14,207][ceph_deploy.cli][INFO ] func : -[2020-03-03 21:49:14,207][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 21:49:14,207][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 21:49:14,207][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts riot01:rgw.riot01 -[2020-03-03 21:49:15,751][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 21:49:15,753][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 21:49:15,923][riot01][DEBUG ] detect machine type -[2020-03-03 21:49:15,964][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-03 21:49:15,965][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-03 21:49:15,966][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to riot01 -[2020-03-03 21:49:15,967][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 21:49:15,995][riot01][DEBUG ] create path recursively if it doesn't exist -[2020-03-03 21:49:16,010][riot01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.riot01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.riot01/keyring -[2020-03-03 21:49:16,891][riot01][ERROR ] Traceback (most recent call last): -[2020-03-03 21:49:16,892][riot01][ERROR ] File "/usr/bin/ceph", line 1266, in -[2020-03-03 21:49:16,892][riot01][ERROR ] retval = main() -[2020-03-03 21:49:16,892][riot01][ERROR ] File "/usr/bin/ceph", line 979, in main -[2020-03-03 21:49:16,892][riot01][ERROR ] conffile=conffile) -[2020-03-03 21:49:16,893][riot01][ERROR ] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-03 21:49:16,893][riot01][ERROR ] raise Exception("timed out") -[2020-03-03 21:49:16,893][riot01][ERROR ] Exception: timed out -[2020-03-03 21:49:16,893][riot01][ERROR ] exit code from command was: 1 -[2020-03-03 21:49:16,893][ceph_deploy.rgw][ERROR ] could not create rgw -[2020-03-03 21:49:16,894][ceph_deploy][ERROR ] GenericError: Failed to create 1 RGWs - -[2020-03-03 22:03:53,501][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin lenny -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] username : root -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] client : ['lenny'] -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] func : -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 22:03:53,502][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 22:03:53,502][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to lenny -[2020-03-03 22:03:53,842][lenny][DEBUG ] connected to host: root@lenny -[2020-03-03 22:03:53,843][lenny][DEBUG ] detect platform information from remote host -[2020-03-03 22:03:53,859][lenny][DEBUG ] detect machine type -[2020-03-03 22:03:53,862][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:03:53,864][ceph_deploy.admin][ERROR ] RuntimeError: config file /etc/ceph/ceph.conf exists with different content; use --overwrite-conf to overwrite -[2020-03-03 22:03:53,864][ceph_deploy][ERROR ] GenericError: Failed to configure 1 admin hosts - -[2020-03-03 22:04:06,808][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 22:04:06,808][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push lenny -[2020-03-03 22:04:06,808][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] username : root -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] client : ['lenny'] -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] func : -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 22:04:06,809][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 22:04:06,809][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-03 22:04:07,221][lenny][DEBUG ] connected to host: root@lenny -[2020-03-03 22:04:07,221][lenny][DEBUG ] detect platform information from remote host -[2020-03-03 22:04:07,236][lenny][DEBUG ] detect machine type -[2020-03-03 22:04:07,240][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:04:29,625][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin lenny -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] username : root -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] client : ['lenny'] -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] func : -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 22:04:29,625][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 22:04:29,625][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to lenny -[2020-03-03 22:04:29,982][lenny][DEBUG ] connected to host: root@lenny -[2020-03-03 22:04:29,982][lenny][DEBUG ] detect platform information from remote host -[2020-03-03 22:04:29,997][lenny][DEBUG ] detect machine type -[2020-03-03 22:04:30,000][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:26:06,567][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] username : root -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] verbose : False -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] quiet : False -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] func : -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-03 22:26:06,568][ceph_deploy.cli][INFO ] default_release : False -[2020-03-03 22:26:06,568][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-03 22:26:08,221][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-03 22:26:08,223][riot01][DEBUG ] detect platform information from remote host -[2020-03-03 22:26:08,397][riot01][DEBUG ] detect machine type -[2020-03-03 22:26:08,442][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:26:08,473][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-03 22:26:11,728][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-03 22:26:11,729][pine01][DEBUG ] detect platform information from remote host -[2020-03-03 22:26:11,936][pine01][DEBUG ] detect machine type -[2020-03-03 22:26:11,963][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:26:11,981][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-03 22:26:13,617][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-03 22:26:13,619][pine02][DEBUG ] detect platform information from remote host -[2020-03-03 22:26:13,739][pine02][DEBUG ] detect machine type -[2020-03-03 22:26:13,767][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:26:13,780][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-03 22:26:16,241][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-03 22:26:16,243][ebin01][DEBUG ] detect platform information from remote host -[2020-03-03 22:26:16,458][ebin01][DEBUG ] detect machine type -[2020-03-03 22:26:16,491][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:26:16,508][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-03 22:26:18,915][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-03 22:26:18,916][ebin02][DEBUG ] detect platform information from remote host -[2020-03-03 22:26:19,127][ebin02][DEBUG ] detect machine type -[2020-03-03 22:26:19,167][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-03 22:26:19,182][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-03 22:26:19,651][lenny][DEBUG ] connected to host: root@lenny -[2020-03-03 22:26:19,652][lenny][DEBUG ] detect platform information from remote host -[2020-03-03 22:26:19,666][lenny][DEBUG ] detect machine type -[2020-03-03 22:26:19,670][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 17:33:41,452][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy mon add riot01 -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] username : None -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] func : -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] address : None -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 17:33:41,453][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 17:33:41,454][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 17:33:41,455][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 17:33:44,052][ceph_deploy][ERROR ] KeyboardInterrupt - -[2020-03-04 17:33:50,078][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 17:33:50,078][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-04 17:33:50,078][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] username : root -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] func : -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] address : None -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 17:33:50,079][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 17:33:50,079][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 17:33:50,080][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 17:33:51,705][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 17:33:51,717][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 17:33:51,882][riot01][DEBUG ] detect machine type -[2020-03-04 17:33:51,923][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 17:33:51,950][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 17:33:51,952][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-04 17:33:51,953][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 17:33:53,525][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 17:33:53,527][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 17:33:53,696][riot01][DEBUG ] detect machine type -[2020-03-04 17:33:53,733][riot01][DEBUG ] find the location of an executable -[2020-03-04 17:33:53,739][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 17:33:53,740][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 17:33:53,740][riot01][DEBUG ] get remote short hostname -[2020-03-04 17:33:53,747][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 17:33:53,748][riot01][DEBUG ] get remote short hostname -[2020-03-04 17:33:53,765][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 17:33:53,786][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 17:33:53,798][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 17:33:53,804][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 17:33:53,809][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 17:33:53,810][riot01][DEBUG ] create the monitor keyring file -[2020-03-04 17:33:53,829][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 17:33:54,877][riot01][WARNING] Traceback (most recent call last): -[2020-03-04 17:33:54,877][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-04 17:33:54,887][riot01][WARNING] retval = main() -[2020-03-04 17:33:54,888][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-04 17:33:54,888][riot01][WARNING] conffile=conffile) -[2020-03-04 17:33:54,888][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-04 17:33:54,888][riot01][WARNING] raise Exception("timed out") -[2020-03-04 17:33:54,889][riot01][WARNING] Exception: timed out -[2020-03-04 17:33:55,004][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-04 17:33:55,005][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 17:33:55,005][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-04 20:18:17,335][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:18:17,335][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-04 20:18:17,335][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:18:17,335][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:18:17,335][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:18:17,335][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] address : None -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:18:17,336][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:18:17,336][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 20:18:17,336][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 20:18:18,819][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:18:18,820][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:18:18,996][riot01][DEBUG ] detect machine type -[2020-03-04 20:18:19,030][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:18:19,058][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 20:18:19,060][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-04 20:18:19,061][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:18:20,610][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:18:20,611][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:18:20,781][riot01][DEBUG ] detect machine type -[2020-03-04 20:18:20,816][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:18:20,821][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:18:20,821][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:18:20,821][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:18:20,827][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 20:18:20,827][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:18:20,842][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:18:20,862][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:18:20,869][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:18:20,873][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:18:20,878][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 20:18:20,878][riot01][DEBUG ] create the monitor keyring file -[2020-03-04 20:18:20,892][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:18:21,663][riot01][WARNING] Traceback (most recent call last): -[2020-03-04 20:18:21,663][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-04 20:18:21,663][riot01][WARNING] retval = main() -[2020-03-04 20:18:21,663][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-04 20:18:21,663][riot01][WARNING] conffile=conffile) -[2020-03-04 20:18:21,663][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-04 20:18:21,664][riot01][WARNING] raise Exception("timed out") -[2020-03-04 20:18:21,664][riot01][WARNING] Exception: timed out -[2020-03-04 20:18:21,777][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-04 20:18:21,778][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:18:21,778][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-04 20:18:35,097][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purge riot01 -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] host : ['riot01'] -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:18:35,097][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:18:35,098][ceph_deploy.install][INFO ] note that some dependencies *will not* be removed because they can cause issues with qemu-kvm -[2020-03-04 20:18:35,098][ceph_deploy.install][INFO ] like: librbd1 and librados2 -[2020-03-04 20:18:35,098][ceph_deploy.install][DEBUG ] Purging on cluster ceph hosts riot01 -[2020-03-04 20:18:35,098][ceph_deploy.install][DEBUG ] Detecting platform for host riot01 ... -[2020-03-04 20:18:36,581][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:18:36,581][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:18:36,752][riot01][DEBUG ] detect machine type -[2020-03-04 20:18:36,787][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-04 20:18:36,787][riot01][INFO ] Purging Ceph on riot01 -[2020-03-04 20:18:36,797][riot01][INFO ] Running command: env DEBIAN_FRONTEND=noninteractive DEBIAN_PRIORITY=critical apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw -[2020-03-04 20:18:37,316][riot01][DEBUG ] Reading package lists... -[2020-03-04 20:18:39,284][riot01][DEBUG ] Building dependency tree... -[2020-03-04 20:18:39,284][riot01][DEBUG ] Reading state information... -[2020-03-04 20:18:41,051][riot01][DEBUG ] Package 'ceph-fs-common' is not installed, so not removed -[2020-03-04 20:18:41,051][riot01][DEBUG ] The following packages were automatically installed and are no longer required: -[2020-03-04 20:18:41,051][riot01][DEBUG ] cryptsetup-bin dmeventd gdisk hdparm libaio1 libbabeltrace1 -[2020-03-04 20:18:41,051][riot01][DEBUG ] libboost-date-time1.67.0 libboost-program-options1.67.0 -[2020-03-04 20:18:41,055][riot01][DEBUG ] libboost-python1.67.0 libboost-random1.67.0 libboost-regex1.67.0 libcephfs2 -[2020-03-04 20:18:41,062][riot01][DEBUG ] libdevmapper-event1.02.1 libdw1 libfuse2 libgoogle-perftools4 libleveldb1d -[2020-03-04 20:18:41,065][riot01][DEBUG ] liblvm2cmd2.03 liboath0 libpython3.7 libradosstriper1 librbd1 libreadline5 -[2020-03-04 20:18:41,073][riot01][DEBUG ] libtcmalloc-minimal4 lvm2 python-pastedeploy-tpl python3-asn1crypto -[2020-03-04 20:18:41,076][riot01][DEBUG ] python3-bcrypt python3-bs4 python3-ceph-argparse python3-cephfs -[2020-03-04 20:18:41,091][riot01][DEBUG ] python3-certifi python3-cffi-backend python3-chardet python3-cherrypy3 -[2020-03-04 20:18:41,092][riot01][DEBUG ] python3-cryptography python3-idna python3-jwt python3-logutils python3-mako -[2020-03-04 20:18:41,092][riot01][DEBUG ] python3-markupsafe python3-openssl python3-paste python3-pastedeploy -[2020-03-04 20:18:41,095][riot01][DEBUG ] python3-pecan python3-pkg-resources python3-prettytable python3-rados -[2020-03-04 20:18:41,096][riot01][DEBUG ] python3-rbd python3-requests python3-simplegeneric python3-singledispatch -[2020-03-04 20:18:41,099][riot01][DEBUG ] python3-soupsieve python3-tempita python3-urllib3 python3-waitress -[2020-03-04 20:18:41,101][riot01][DEBUG ] python3-webob python3-webtest python3-werkzeug xfsprogs -[2020-03-04 20:18:41,104][riot01][DEBUG ] Use 'apt autoremove' to remove them. -[2020-03-04 20:18:41,719][riot01][DEBUG ] The following packages will be REMOVED: -[2020-03-04 20:18:41,719][riot01][DEBUG ] ceph* ceph-base* ceph-common* ceph-mds* ceph-mgr* ceph-mon* ceph-osd* -[2020-03-04 20:18:41,719][riot01][DEBUG ] radosgw* -[2020-03-04 20:18:45,339][riot01][DEBUG ] 0 upgraded, 0 newly installed, 8 to remove and 3 not upgraded. -[2020-03-04 20:18:45,339][riot01][DEBUG ] After this operation, 168 MB disk space will be freed. -[2020-03-04 20:18:46,905][riot01][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 58434 files and directories currently installed.) -[2020-03-04 20:18:46,937][riot01][DEBUG ] Removing ceph-mds (14.2.7-1~bpo10+1) ... -[2020-03-04 20:18:53,211][riot01][DEBUG ] Removing ceph (14.2.7-1~bpo10+1) ... -[2020-03-04 20:18:53,575][riot01][DEBUG ] Removing ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:01,605][riot01][DEBUG ] Removing ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:07,830][riot01][DEBUG ] Removing radosgw (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:16,962][riot01][DEBUG ] Removing ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:22,184][riot01][DEBUG ] Removing ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:27,759][riot01][DEBUG ] Removing ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:35,084][riot01][DEBUG ] Processing triggers for libc-bin (2.28-10) ... -[2020-03-04 20:19:38,303][riot01][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 58016 files and directories currently installed.) -[2020-03-04 20:19:38,303][riot01][DEBUG ] Purging configuration files for radosgw (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:47,883][riot01][DEBUG ] dpkg: warning: while removing radosgw, directory '/var/lib/ceph/radosgw' not empty so not removed -[2020-03-04 20:19:47,997][riot01][DEBUG ] Purging configuration files for ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-04 20:19:55,275][riot01][DEBUG ] dpkg: warning: while removing ceph-mon, directory '/var/lib/ceph/mon' not empty so not removed -[2020-03-04 20:19:55,388][riot01][DEBUG ] Purging configuration files for ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-04 20:20:02,716][riot01][DEBUG ] dpkg: warning: while removing ceph-base, directory '/var/lib/ceph/tmp' not empty so not removed -[2020-03-04 20:20:02,717][riot01][DEBUG ] dpkg: warning: while removing ceph-base, directory '/var/lib/ceph/bootstrap-rgw' not empty so not removed -[2020-03-04 20:20:02,717][riot01][DEBUG ] dpkg: warning: while removing ceph-base, directory '/var/lib/ceph/bootstrap-mgr' not empty so not removed -[2020-03-04 20:20:02,830][riot01][DEBUG ] Purging configuration files for ceph-mds (14.2.7-1~bpo10+1) ... -[2020-03-04 20:20:10,208][riot01][DEBUG ] Purging configuration files for ceph (14.2.7-1~bpo10+1) ... -[2020-03-04 20:20:10,522][riot01][DEBUG ] Purging configuration files for ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-04 20:20:18,751][riot01][DEBUG ] dpkg: warning: while removing ceph-mgr, directory '/var/lib/ceph/mgr' not empty so not removed -[2020-03-04 20:20:18,865][riot01][DEBUG ] Purging configuration files for ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-04 20:20:28,809][riot01][DEBUG ] Purging configuration files for ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-04 20:20:38,858][riot01][DEBUG ] Processing triggers for systemd (241-7~deb10u3) ... -[2020-03-04 20:20:46,650][riot01][DEBUG ] [master 22acdd4] committing changes in /etc made by "apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw" -[2020-03-04 20:20:46,651][riot01][DEBUG ] 41 files changed, 2 insertions(+), 293 deletions(-) -[2020-03-04 20:20:46,651][riot01][DEBUG ] delete mode 100644 ceph/ceph.client.admin.keyring -[2020-03-04 20:20:46,651][riot01][DEBUG ] delete mode 100644 ceph/ceph.conf -[2020-03-04 20:20:46,651][riot01][DEBUG ] delete mode 100644 ceph/rbdmap -[2020-03-04 20:20:46,652][riot01][DEBUG ] delete mode 100644 ceph/tmp9xog9V -[2020-03-04 20:20:46,652][riot01][DEBUG ] delete mode 100644 ceph/tmpCK3mc0 -[2020-03-04 20:20:46,652][riot01][DEBUG ] delete mode 100644 default/ceph -[2020-03-04 20:20:46,656][riot01][DEBUG ] delete mode 100755 init.d/radosgw -[2020-03-04 20:20:46,657][riot01][DEBUG ] delete mode 100755 init.d/rbdmap -[2020-03-04 20:20:46,658][riot01][DEBUG ] delete mode 100644 logrotate.d/ceph-common -[2020-03-04 20:20:46,659][riot01][DEBUG ] delete mode 120000 rc0.d/K01radosgw -[2020-03-04 20:20:46,659][riot01][DEBUG ] delete mode 120000 rc0.d/K01rbdmap -[2020-03-04 20:20:46,660][riot01][DEBUG ] delete mode 120000 rc1.d/K01radosgw -[2020-03-04 20:20:46,661][riot01][DEBUG ] delete mode 120000 rc1.d/K01rbdmap -[2020-03-04 20:20:46,661][riot01][DEBUG ] delete mode 120000 rc2.d/S01radosgw -[2020-03-04 20:20:46,662][riot01][DEBUG ] delete mode 120000 rc2.d/S01rbdmap -[2020-03-04 20:20:46,663][riot01][DEBUG ] delete mode 120000 rc3.d/S01radosgw -[2020-03-04 20:20:46,663][riot01][DEBUG ] delete mode 120000 rc3.d/S01rbdmap -[2020-03-04 20:20:46,663][riot01][DEBUG ] delete mode 120000 rc4.d/S01radosgw -[2020-03-04 20:20:46,664][riot01][DEBUG ] delete mode 120000 rc4.d/S01rbdmap -[2020-03-04 20:20:46,667][riot01][DEBUG ] delete mode 120000 rc5.d/S01radosgw -[2020-03-04 20:20:46,669][riot01][DEBUG ] delete mode 120000 rc5.d/S01rbdmap -[2020-03-04 20:20:46,676][riot01][DEBUG ] delete mode 120000 rc6.d/K01radosgw -[2020-03-04 20:20:46,677][riot01][DEBUG ] delete mode 120000 rc6.d/K01rbdmap -[2020-03-04 20:20:46,678][riot01][DEBUG ] delete mode 100644 sudoers.d/ceph-osd-smartctl -[2020-03-04 20:20:46,681][riot01][DEBUG ] delete mode 100644 sysctl.d/30-ceph-osd.conf -[2020-03-04 20:20:46,682][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-crash.service -[2020-03-04 20:20:46,698][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mds.target -[2020-03-04 20:20:46,698][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mgr.target -[2020-03-04 20:20:46,699][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mon.target -[2020-03-04 20:20:46,707][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-osd.target -[2020-03-04 20:20:46,708][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-radosgw.target -[2020-03-04 20:20:46,712][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mds.target -[2020-03-04 20:20:46,714][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mgr.target -[2020-03-04 20:20:46,722][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mon.target -[2020-03-04 20:20:46,722][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-osd.target -[2020-03-04 20:20:46,723][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-radosgw.target -[2020-03-04 20:20:46,724][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph.target -[2020-03-04 20:20:46,733][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/rbdmap.service -[2020-03-04 20:20:46,897][riot01][WARNING] W: --force-yes is deprecated, use one of the options starting with --allow instead. -[2020-03-04 20:21:04,695][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:21:04,695][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purgedata riot01 -[2020-03-04 20:21:04,695][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:21:04,695][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:21:04,695][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:21:04,695][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:21:04,696][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:21:04,696][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:21:04,696][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:21:04,696][ceph_deploy.cli][INFO ] host : ['riot01'] -[2020-03-04 20:21:04,696][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:21:04,696][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:21:04,696][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:21:04,696][ceph_deploy.install][DEBUG ] Purging data from cluster ceph hosts riot01 -[2020-03-04 20:21:06,207][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:21:06,209][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:21:06,376][riot01][DEBUG ] detect machine type -[2020-03-04 20:21:06,416][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:21:07,987][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:21:07,990][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:21:08,157][riot01][DEBUG ] detect machine type -[2020-03-04 20:21:08,198][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-04 20:21:08,199][riot01][INFO ] purging data on riot01 -[2020-03-04 20:21:08,209][riot01][INFO ] Running command: rm -rf --one-file-system -- /var/lib/ceph -[2020-03-04 20:21:08,361][riot01][INFO ] Running command: rm -rf --one-file-system -- /etc/ceph/ -[2020-03-04 20:26:43,678][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] address : None -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:26:43,679][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:26:43,680][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 20:26:43,680][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 20:26:45,195][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:26:45,196][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:26:45,365][riot01][DEBUG ] detect machine type -[2020-03-04 20:26:45,404][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:26:45,429][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 20:26:45,434][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-04 20:26:45,435][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:26:46,995][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:26:46,997][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:26:47,164][riot01][DEBUG ] detect machine type -[2020-03-04 20:26:47,198][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:26:47,203][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:26:47,203][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:26:47,203][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:26:47,208][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 20:26:47,208][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:26:47,224][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:26:47,243][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:26:47,249][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:26:47,256][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:26:47,261][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 20:26:47,261][riot01][DEBUG ] create the monitor keyring file -[2020-03-04 20:26:47,277][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:26:48,049][riot01][WARNING] Traceback (most recent call last): -[2020-03-04 20:26:48,049][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-04 20:26:48,059][riot01][WARNING] retval = main() -[2020-03-04 20:26:48,060][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-04 20:26:48,060][riot01][WARNING] conffile=conffile) -[2020-03-04 20:26:48,060][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-04 20:26:48,061][riot01][WARNING] raise Exception("timed out") -[2020-03-04 20:26:48,062][riot01][WARNING] Exception: timed out -[2020-03-04 20:26:48,177][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-04 20:26:48,177][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:26:48,178][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-04 20:29:14,067][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] address : None -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:29:14,068][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:29:14,069][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 20:29:14,069][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 20:29:15,552][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:29:15,554][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:29:15,724][riot01][DEBUG ] detect machine type -[2020-03-04 20:29:15,761][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:29:15,791][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 20:29:15,795][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-04 20:29:15,796][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:29:17,361][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:29:17,363][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:29:17,533][riot01][DEBUG ] detect machine type -[2020-03-04 20:29:17,575][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:29:17,582][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:29:17,582][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:29:17,583][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:29:17,589][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 20:29:17,590][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:29:17,607][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:29:17,630][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:29:17,637][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:29:17,643][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:29:17,649][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 20:29:17,650][riot01][DEBUG ] create the monitor keyring file -[2020-03-04 20:29:17,671][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:29:18,458][riot01][WARNING] Traceback (most recent call last): -[2020-03-04 20:29:18,459][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-04 20:29:18,459][riot01][WARNING] retval = main() -[2020-03-04 20:29:18,459][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-04 20:29:18,459][riot01][WARNING] conffile=conffile) -[2020-03-04 20:29:18,460][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-04 20:29:18,460][riot01][WARNING] raise Exception("timed out") -[2020-03-04 20:29:18,460][riot01][WARNING] Exception: timed out -[2020-03-04 20:29:18,575][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-04 20:29:18,576][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:29:18,576][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-04 20:30:36,948][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:30:36,948][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] address : None -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:30:36,949][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:30:36,949][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 20:30:36,950][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 20:30:38,442][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:30:38,444][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:30:38,613][riot01][DEBUG ] detect machine type -[2020-03-04 20:30:38,654][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:30:38,683][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 20:30:38,688][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-04 20:30:38,688][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:30:40,338][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:30:40,340][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:30:40,508][riot01][DEBUG ] detect machine type -[2020-03-04 20:30:40,546][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:30:40,552][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:30:40,553][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:30:40,553][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:30:40,559][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 20:30:40,560][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:30:40,577][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:30:40,599][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:30:40,605][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:30:40,610][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:30:40,617][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 20:30:40,617][riot01][DEBUG ] create the monitor keyring file -[2020-03-04 20:30:40,638][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:30:41,423][riot01][WARNING] Traceback (most recent call last): -[2020-03-04 20:30:41,424][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-04 20:30:41,424][riot01][WARNING] retval = main() -[2020-03-04 20:30:41,424][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-04 20:30:41,424][riot01][WARNING] conffile=conffile) -[2020-03-04 20:30:41,425][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-04 20:30:41,425][riot01][WARNING] raise Exception("timed out") -[2020-03-04 20:30:41,425][riot01][WARNING] Exception: timed out -[2020-03-04 20:30:41,540][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-04 20:30:41,540][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:30:41,540][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-04 20:31:01,222][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] address : None -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:31:01,222][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:31:01,223][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 20:31:01,223][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 20:31:02,796][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:31:02,798][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:31:02,970][riot01][DEBUG ] detect machine type -[2020-03-04 20:31:03,015][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:31:03,058][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 20:31:03,062][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-04 20:31:03,062][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:31:04,628][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:31:04,630][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:31:04,799][riot01][DEBUG ] detect machine type -[2020-03-04 20:31:04,836][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:31:04,842][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:31:04,842][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:31:04,843][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:31:04,849][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 20:31:04,850][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:31:04,868][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:31:04,890][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:31:04,896][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:31:04,902][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:31:04,909][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 20:31:04,909][riot01][DEBUG ] create the monitor keyring file -[2020-03-04 20:31:04,928][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:31:05,716][riot01][WARNING] Traceback (most recent call last): -[2020-03-04 20:31:05,717][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-04 20:31:05,717][riot01][WARNING] retval = main() -[2020-03-04 20:31:05,718][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-04 20:31:05,718][riot01][WARNING] conffile=conffile) -[2020-03-04 20:31:05,718][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-04 20:31:05,719][riot01][WARNING] raise Exception("timed out") -[2020-03-04 20:31:05,719][riot01][WARNING] Exception: timed out -[2020-03-04 20:31:05,834][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-04 20:31:05,834][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-04 20:31:05,835][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-04 20:34:03,859][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:34:03,859][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon create riot01 -[2020-03-04 20:34:03,859][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:34:03,859][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:34:03,859][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] keyrings : None -[2020-03-04 20:34:03,860][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:34:03,860][ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts riot01 -[2020-03-04 20:34:03,860][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:34:05,359][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:34:05,361][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:34:05,531][riot01][DEBUG ] detect machine type -[2020-03-04 20:34:05,573][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:34:05,580][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:34:05,581][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:34:05,581][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:34:05,588][riot01][DEBUG ] deploying mon to riot01 -[2020-03-04 20:34:05,588][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:34:05,594][riot01][DEBUG ] remote hostname: riot01 -[2020-03-04 20:34:05,606][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:34:05,628][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:34:05,634][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:34:05,640][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:34:05,646][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 20:34:05,646][riot01][DEBUG ] create the monitor keyring file -[2020-03-04 20:34:05,663][riot01][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i riot01 --keyring /var/lib/ceph/tmp/ceph-riot01.mon.keyring --setuser 64045 --setgroup 64045 -[2020-03-04 20:34:06,398][riot01][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-04 20:34:06,405][riot01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-04 20:34:06,411][riot01][DEBUG ] create the init path if it does not exist -[2020-03-04 20:34:06,431][riot01][INFO ] Running command: systemctl enable ceph.target -[2020-03-04 20:34:08,678][riot01][INFO ] Running command: systemctl enable ceph-mon@riot01 -[2020-03-04 20:34:10,921][riot01][INFO ] Running command: systemctl start ceph-mon@riot01 -[2020-03-04 20:34:13,114][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-04 20:34:14,033][riot01][DEBUG ] ******************************************************************************** -[2020-03-04 20:34:14,034][riot01][DEBUG ] status for monitor: mon.riot01 -[2020-03-04 20:34:14,035][riot01][DEBUG ] { -[2020-03-04 20:34:14,035][riot01][DEBUG ] "election_epoch": 0, -[2020-03-04 20:34:14,035][riot01][DEBUG ] "extra_probe_peers": [ -[2020-03-04 20:34:14,036][riot01][DEBUG ] { -[2020-03-04 20:34:14,036][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:14,036][riot01][DEBUG ] { -[2020-03-04 20:34:14,036][riot01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-04 20:34:14,037][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:14,037][riot01][DEBUG ] "type": "v2" -[2020-03-04 20:34:14,037][riot01][DEBUG ] }, -[2020-03-04 20:34:14,037][riot01][DEBUG ] { -[2020-03-04 20:34:14,038][riot01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-04 20:34:14,038][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:14,038][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:14,038][riot01][DEBUG ] } -[2020-03-04 20:34:14,039][riot01][DEBUG ] ] -[2020-03-04 20:34:14,039][riot01][DEBUG ] }, -[2020-03-04 20:34:14,039][riot01][DEBUG ] { -[2020-03-04 20:34:14,039][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:14,040][riot01][DEBUG ] { -[2020-03-04 20:34:14,040][riot01][DEBUG ] "addr": "192.168.10.161:3300", -[2020-03-04 20:34:14,040][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:14,040][riot01][DEBUG ] "type": "v2" -[2020-03-04 20:34:14,041][riot01][DEBUG ] }, -[2020-03-04 20:34:14,041][riot01][DEBUG ] { -[2020-03-04 20:34:14,041][riot01][DEBUG ] "addr": "192.168.10.161:6789", -[2020-03-04 20:34:14,041][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:14,042][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:14,042][riot01][DEBUG ] } -[2020-03-04 20:34:14,042][riot01][DEBUG ] ] -[2020-03-04 20:34:14,042][riot01][DEBUG ] } -[2020-03-04 20:34:14,043][riot01][DEBUG ] ], -[2020-03-04 20:34:14,043][riot01][DEBUG ] "feature_map": { -[2020-03-04 20:34:14,043][riot01][DEBUG ] "mon": [ -[2020-03-04 20:34:14,043][riot01][DEBUG ] { -[2020-03-04 20:34:14,044][riot01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-04 20:34:14,044][riot01][DEBUG ] "num": 1, -[2020-03-04 20:34:14,044][riot01][DEBUG ] "release": "luminous" -[2020-03-04 20:34:14,044][riot01][DEBUG ] } -[2020-03-04 20:34:14,045][riot01][DEBUG ] ] -[2020-03-04 20:34:14,045][riot01][DEBUG ] }, -[2020-03-04 20:34:14,045][riot01][DEBUG ] "features": { -[2020-03-04 20:34:14,046][riot01][DEBUG ] "quorum_con": "0", -[2020-03-04 20:34:14,046][riot01][DEBUG ] "quorum_mon": [], -[2020-03-04 20:34:14,046][riot01][DEBUG ] "required_con": "0", -[2020-03-04 20:34:14,047][riot01][DEBUG ] "required_mon": [] -[2020-03-04 20:34:14,047][riot01][DEBUG ] }, -[2020-03-04 20:34:14,047][riot01][DEBUG ] "monmap": { -[2020-03-04 20:34:14,047][riot01][DEBUG ] "created": "2020-03-04 20:34:06.060053", -[2020-03-04 20:34:14,048][riot01][DEBUG ] "epoch": 0, -[2020-03-04 20:34:14,048][riot01][DEBUG ] "features": { -[2020-03-04 20:34:14,048][riot01][DEBUG ] "optional": [], -[2020-03-04 20:34:14,048][riot01][DEBUG ] "persistent": [] -[2020-03-04 20:34:14,049][riot01][DEBUG ] }, -[2020-03-04 20:34:14,049][riot01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-04 20:34:14,049][riot01][DEBUG ] "min_mon_release": 0, -[2020-03-04 20:34:14,049][riot01][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-04 20:34:14,050][riot01][DEBUG ] "modified": "2020-03-04 20:34:06.060053", -[2020-03-04 20:34:14,050][riot01][DEBUG ] "mons": [ -[2020-03-04 20:34:14,050][riot01][DEBUG ] { -[2020-03-04 20:34:14,050][riot01][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-04 20:34:14,051][riot01][DEBUG ] "name": "pine01", -[2020-03-04 20:34:14,051][riot01][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-04 20:34:14,051][riot01][DEBUG ] "public_addrs": { -[2020-03-04 20:34:14,051][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:14,052][riot01][DEBUG ] { -[2020-03-04 20:34:14,052][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-04 20:34:14,052][riot01][DEBUG ] "nonce": 1, -[2020-03-04 20:34:14,052][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:14,052][riot01][DEBUG ] } -[2020-03-04 20:34:14,053][riot01][DEBUG ] ] -[2020-03-04 20:34:14,053][riot01][DEBUG ] }, -[2020-03-04 20:34:14,053][riot01][DEBUG ] "rank": 0 -[2020-03-04 20:34:14,053][riot01][DEBUG ] }, -[2020-03-04 20:34:14,054][riot01][DEBUG ] { -[2020-03-04 20:34:14,054][riot01][DEBUG ] "addr": "0.0.0.0:0/2", -[2020-03-04 20:34:14,054][riot01][DEBUG ] "name": "pine02", -[2020-03-04 20:34:14,054][riot01][DEBUG ] "public_addr": "0.0.0.0:0/2", -[2020-03-04 20:34:14,055][riot01][DEBUG ] "public_addrs": { -[2020-03-04 20:34:14,055][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:14,055][riot01][DEBUG ] { -[2020-03-04 20:34:14,055][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-04 20:34:14,056][riot01][DEBUG ] "nonce": 2, -[2020-03-04 20:34:14,056][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:14,056][riot01][DEBUG ] } -[2020-03-04 20:34:14,056][riot01][DEBUG ] ] -[2020-03-04 20:34:14,057][riot01][DEBUG ] }, -[2020-03-04 20:34:14,057][riot01][DEBUG ] "rank": 1 -[2020-03-04 20:34:14,057][riot01][DEBUG ] } -[2020-03-04 20:34:14,057][riot01][DEBUG ] ] -[2020-03-04 20:34:14,058][riot01][DEBUG ] }, -[2020-03-04 20:34:14,058][riot01][DEBUG ] "name": "riot01", -[2020-03-04 20:34:14,058][riot01][DEBUG ] "outside_quorum": [], -[2020-03-04 20:34:14,058][riot01][DEBUG ] "quorum": [], -[2020-03-04 20:34:14,059][riot01][DEBUG ] "rank": -1, -[2020-03-04 20:34:14,059][riot01][DEBUG ] "state": "probing", -[2020-03-04 20:34:14,059][riot01][DEBUG ] "sync_provider": [] -[2020-03-04 20:34:14,059][riot01][DEBUG ] } -[2020-03-04 20:34:14,060][riot01][DEBUG ] ******************************************************************************** -[2020-03-04 20:34:14,060][riot01][INFO ] monitor: mon.riot01 is currently at the state of probing -[2020-03-04 20:34:14,074][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-04 20:34:14,946][riot01][WARNING] riot01 is not defined in `mon initial members` -[2020-03-04 20:34:14,947][riot01][WARNING] monitor riot01 does not exist in monmap -[2020-03-04 20:34:23,993][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:34:23,993][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-04 20:34:23,993][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:34:23,993][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] address : None -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:34:23,994][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:34:23,994][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 20:34:23,994][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 20:34:25,511][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:34:25,514][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:34:25,682][riot01][DEBUG ] detect machine type -[2020-03-04 20:34:25,720][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:34:25,750][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 20:34:25,754][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-04 20:34:25,755][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:34:27,882][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:34:27,884][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:34:28,603][riot01][DEBUG ] detect machine type -[2020-03-04 20:34:28,740][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:34:28,747][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:34:28,748][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:34:28,748][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:34:28,757][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 20:34:28,757][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:34:28,774][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:34:28,812][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:34:28,826][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:34:28,833][riot01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-04 20:34:28,839][riot01][DEBUG ] create the init path if it does not exist -[2020-03-04 20:34:28,864][riot01][INFO ] Running command: systemctl enable ceph.target -[2020-03-04 20:34:31,332][riot01][INFO ] Running command: systemctl enable ceph-mon@riot01 -[2020-03-04 20:34:33,733][riot01][INFO ] Running command: systemctl start ceph-mon@riot01 -[2020-03-04 20:34:35,972][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-04 20:34:36,845][riot01][WARNING] riot01 is not defined in `mon initial members` -[2020-03-04 20:34:36,845][riot01][WARNING] monitor riot01 does not exist in monmap -[2020-03-04 20:34:36,860][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-04 20:34:37,731][riot01][DEBUG ] ******************************************************************************** -[2020-03-04 20:34:37,731][riot01][DEBUG ] status for monitor: mon.riot01 -[2020-03-04 20:34:37,732][riot01][DEBUG ] { -[2020-03-04 20:34:37,732][riot01][DEBUG ] "election_epoch": 0, -[2020-03-04 20:34:37,733][riot01][DEBUG ] "extra_probe_peers": [ -[2020-03-04 20:34:37,733][riot01][DEBUG ] { -[2020-03-04 20:34:37,733][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:37,733][riot01][DEBUG ] { -[2020-03-04 20:34:37,733][riot01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-04 20:34:37,734][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:37,734][riot01][DEBUG ] "type": "v2" -[2020-03-04 20:34:37,734][riot01][DEBUG ] }, -[2020-03-04 20:34:37,734][riot01][DEBUG ] { -[2020-03-04 20:34:37,734][riot01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-04 20:34:37,734][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:37,735][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:37,735][riot01][DEBUG ] } -[2020-03-04 20:34:37,735][riot01][DEBUG ] ] -[2020-03-04 20:34:37,735][riot01][DEBUG ] }, -[2020-03-04 20:34:37,735][riot01][DEBUG ] { -[2020-03-04 20:34:37,735][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:37,736][riot01][DEBUG ] { -[2020-03-04 20:34:37,736][riot01][DEBUG ] "addr": "192.168.10.161:3300", -[2020-03-04 20:34:37,736][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:37,736][riot01][DEBUG ] "type": "v2" -[2020-03-04 20:34:37,736][riot01][DEBUG ] }, -[2020-03-04 20:34:37,736][riot01][DEBUG ] { -[2020-03-04 20:34:37,736][riot01][DEBUG ] "addr": "192.168.10.161:6789", -[2020-03-04 20:34:37,736][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:34:37,736][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:37,736][riot01][DEBUG ] } -[2020-03-04 20:34:37,736][riot01][DEBUG ] ] -[2020-03-04 20:34:37,736][riot01][DEBUG ] } -[2020-03-04 20:34:37,736][riot01][DEBUG ] ], -[2020-03-04 20:34:37,736][riot01][DEBUG ] "feature_map": { -[2020-03-04 20:34:37,737][riot01][DEBUG ] "mon": [ -[2020-03-04 20:34:37,737][riot01][DEBUG ] { -[2020-03-04 20:34:37,737][riot01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-04 20:34:37,737][riot01][DEBUG ] "num": 1, -[2020-03-04 20:34:37,737][riot01][DEBUG ] "release": "luminous" -[2020-03-04 20:34:37,737][riot01][DEBUG ] } -[2020-03-04 20:34:37,737][riot01][DEBUG ] ] -[2020-03-04 20:34:37,737][riot01][DEBUG ] }, -[2020-03-04 20:34:37,737][riot01][DEBUG ] "features": { -[2020-03-04 20:34:37,737][riot01][DEBUG ] "quorum_con": "0", -[2020-03-04 20:34:37,737][riot01][DEBUG ] "quorum_mon": [], -[2020-03-04 20:34:37,737][riot01][DEBUG ] "required_con": "0", -[2020-03-04 20:34:37,737][riot01][DEBUG ] "required_mon": [] -[2020-03-04 20:34:37,737][riot01][DEBUG ] }, -[2020-03-04 20:34:37,737][riot01][DEBUG ] "monmap": { -[2020-03-04 20:34:37,737][riot01][DEBUG ] "created": "2020-03-04 20:34:06.060053", -[2020-03-04 20:34:37,737][riot01][DEBUG ] "epoch": 0, -[2020-03-04 20:34:37,737][riot01][DEBUG ] "features": { -[2020-03-04 20:34:37,737][riot01][DEBUG ] "optional": [], -[2020-03-04 20:34:37,737][riot01][DEBUG ] "persistent": [] -[2020-03-04 20:34:37,738][riot01][DEBUG ] }, -[2020-03-04 20:34:37,738][riot01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-04 20:34:37,738][riot01][DEBUG ] "min_mon_release": 0, -[2020-03-04 20:34:37,738][riot01][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-04 20:34:37,738][riot01][DEBUG ] "modified": "2020-03-04 20:34:06.060053", -[2020-03-04 20:34:37,738][riot01][DEBUG ] "mons": [ -[2020-03-04 20:34:37,738][riot01][DEBUG ] { -[2020-03-04 20:34:37,738][riot01][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-04 20:34:37,738][riot01][DEBUG ] "name": "pine01", -[2020-03-04 20:34:37,738][riot01][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-04 20:34:37,738][riot01][DEBUG ] "public_addrs": { -[2020-03-04 20:34:37,738][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:37,738][riot01][DEBUG ] { -[2020-03-04 20:34:37,738][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-04 20:34:37,738][riot01][DEBUG ] "nonce": 1, -[2020-03-04 20:34:37,738][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:37,738][riot01][DEBUG ] } -[2020-03-04 20:34:37,738][riot01][DEBUG ] ] -[2020-03-04 20:34:37,738][riot01][DEBUG ] }, -[2020-03-04 20:34:37,739][riot01][DEBUG ] "rank": 0 -[2020-03-04 20:34:37,739][riot01][DEBUG ] }, -[2020-03-04 20:34:37,739][riot01][DEBUG ] { -[2020-03-04 20:34:37,739][riot01][DEBUG ] "addr": "0.0.0.0:0/2", -[2020-03-04 20:34:37,739][riot01][DEBUG ] "name": "pine02", -[2020-03-04 20:34:37,739][riot01][DEBUG ] "public_addr": "0.0.0.0:0/2", -[2020-03-04 20:34:37,739][riot01][DEBUG ] "public_addrs": { -[2020-03-04 20:34:37,739][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:34:37,739][riot01][DEBUG ] { -[2020-03-04 20:34:37,739][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-04 20:34:37,739][riot01][DEBUG ] "nonce": 2, -[2020-03-04 20:34:37,739][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:34:37,739][riot01][DEBUG ] } -[2020-03-04 20:34:37,739][riot01][DEBUG ] ] -[2020-03-04 20:34:37,739][riot01][DEBUG ] }, -[2020-03-04 20:34:37,739][riot01][DEBUG ] "rank": 1 -[2020-03-04 20:34:37,739][riot01][DEBUG ] } -[2020-03-04 20:34:37,739][riot01][DEBUG ] ] -[2020-03-04 20:34:37,740][riot01][DEBUG ] }, -[2020-03-04 20:34:37,740][riot01][DEBUG ] "name": "riot01", -[2020-03-04 20:34:37,740][riot01][DEBUG ] "outside_quorum": [], -[2020-03-04 20:34:37,740][riot01][DEBUG ] "quorum": [], -[2020-03-04 20:34:37,740][riot01][DEBUG ] "rank": -1, -[2020-03-04 20:34:37,740][riot01][DEBUG ] "state": "probing", -[2020-03-04 20:34:37,740][riot01][DEBUG ] "sync_provider": [] -[2020-03-04 20:34:37,740][riot01][DEBUG ] } -[2020-03-04 20:34:37,740][riot01][DEBUG ] ******************************************************************************** -[2020-03-04 20:34:37,740][riot01][INFO ] monitor: mon.riot01 is currently at the state of probing -[2020-03-04 20:38:26,521][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:38:26,521][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add --address 192.168.10.164 riot01 -[2020-03-04 20:38:26,521][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:38:26,521][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:38:26,521][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:38:26,521][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-04 20:38:26,521][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-04 20:38:26,521][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:38:26,522][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:38:26,522][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:38:26,522][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-04 20:38:26,522][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:38:26,522][ceph_deploy.cli][INFO ] address : 192.168.10.164 -[2020-03-04 20:38:26,522][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:38:26,522][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:38:26,522][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-04 20:38:26,522][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-04 20:38:29,554][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:38:29,556][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:38:29,726][riot01][DEBUG ] detect machine type -[2020-03-04 20:38:29,766][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:38:29,797][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-04 20:38:29,797][ceph_deploy.mon][DEBUG ] using mon address via --address 192.168.10.164 -[2020-03-04 20:38:29,798][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-04 20:38:31,340][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:38:31,342][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:38:31,511][riot01][DEBUG ] detect machine type -[2020-03-04 20:38:31,549][riot01][DEBUG ] find the location of an executable -[2020-03-04 20:38:31,557][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-04 20:38:31,557][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-04 20:38:31,558][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:38:31,564][riot01][DEBUG ] adding mon to riot01 -[2020-03-04 20:38:31,564][riot01][DEBUG ] get remote short hostname -[2020-03-04 20:38:31,582][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:38:31,604][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-04 20:38:31,610][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-04 20:38:31,616][riot01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-04 20:38:31,622][riot01][DEBUG ] create the init path if it does not exist -[2020-03-04 20:38:31,638][riot01][INFO ] Running command: systemctl enable ceph.target -[2020-03-04 20:38:33,941][riot01][INFO ] Running command: systemctl enable ceph-mon@riot01 -[2020-03-04 20:38:36,182][riot01][INFO ] Running command: systemctl start ceph-mon@riot01 -[2020-03-04 20:38:38,319][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-04 20:38:39,275][riot01][WARNING] riot01 is not defined in `mon initial members` -[2020-03-04 20:38:39,275][riot01][WARNING] monitor riot01 does not exist in monmap -[2020-03-04 20:38:39,289][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-04 20:38:40,161][riot01][DEBUG ] ******************************************************************************** -[2020-03-04 20:38:40,161][riot01][DEBUG ] status for monitor: mon.riot01 -[2020-03-04 20:38:40,162][riot01][DEBUG ] { -[2020-03-04 20:38:40,162][riot01][DEBUG ] "election_epoch": 0, -[2020-03-04 20:38:40,163][riot01][DEBUG ] "extra_probe_peers": [ -[2020-03-04 20:38:40,163][riot01][DEBUG ] { -[2020-03-04 20:38:40,163][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:38:40,163][riot01][DEBUG ] { -[2020-03-04 20:38:40,163][riot01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-04 20:38:40,163][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:38:40,164][riot01][DEBUG ] "type": "v2" -[2020-03-04 20:38:40,164][riot01][DEBUG ] }, -[2020-03-04 20:38:40,164][riot01][DEBUG ] { -[2020-03-04 20:38:40,164][riot01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-04 20:38:40,164][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:38:40,164][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:38:40,164][riot01][DEBUG ] } -[2020-03-04 20:38:40,165][riot01][DEBUG ] ] -[2020-03-04 20:38:40,165][riot01][DEBUG ] }, -[2020-03-04 20:38:40,165][riot01][DEBUG ] { -[2020-03-04 20:38:40,165][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:38:40,165][riot01][DEBUG ] { -[2020-03-04 20:38:40,165][riot01][DEBUG ] "addr": "192.168.10.161:3300", -[2020-03-04 20:38:40,165][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:38:40,166][riot01][DEBUG ] "type": "v2" -[2020-03-04 20:38:40,166][riot01][DEBUG ] }, -[2020-03-04 20:38:40,166][riot01][DEBUG ] { -[2020-03-04 20:38:40,166][riot01][DEBUG ] "addr": "192.168.10.161:6789", -[2020-03-04 20:38:40,166][riot01][DEBUG ] "nonce": 0, -[2020-03-04 20:38:40,166][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:38:40,167][riot01][DEBUG ] } -[2020-03-04 20:38:40,167][riot01][DEBUG ] ] -[2020-03-04 20:38:40,167][riot01][DEBUG ] } -[2020-03-04 20:38:40,167][riot01][DEBUG ] ], -[2020-03-04 20:38:40,167][riot01][DEBUG ] "feature_map": { -[2020-03-04 20:38:40,167][riot01][DEBUG ] "mon": [ -[2020-03-04 20:38:40,167][riot01][DEBUG ] { -[2020-03-04 20:38:40,168][riot01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-04 20:38:40,168][riot01][DEBUG ] "num": 1, -[2020-03-04 20:38:40,168][riot01][DEBUG ] "release": "luminous" -[2020-03-04 20:38:40,168][riot01][DEBUG ] } -[2020-03-04 20:38:40,168][riot01][DEBUG ] ] -[2020-03-04 20:38:40,168][riot01][DEBUG ] }, -[2020-03-04 20:38:40,168][riot01][DEBUG ] "features": { -[2020-03-04 20:38:40,169][riot01][DEBUG ] "quorum_con": "0", -[2020-03-04 20:38:40,169][riot01][DEBUG ] "quorum_mon": [], -[2020-03-04 20:38:40,169][riot01][DEBUG ] "required_con": "0", -[2020-03-04 20:38:40,169][riot01][DEBUG ] "required_mon": [] -[2020-03-04 20:38:40,169][riot01][DEBUG ] }, -[2020-03-04 20:38:40,169][riot01][DEBUG ] "monmap": { -[2020-03-04 20:38:40,169][riot01][DEBUG ] "created": "2020-03-04 20:34:06.060053", -[2020-03-04 20:38:40,170][riot01][DEBUG ] "epoch": 0, -[2020-03-04 20:38:40,170][riot01][DEBUG ] "features": { -[2020-03-04 20:38:40,170][riot01][DEBUG ] "optional": [], -[2020-03-04 20:38:40,170][riot01][DEBUG ] "persistent": [] -[2020-03-04 20:38:40,170][riot01][DEBUG ] }, -[2020-03-04 20:38:40,170][riot01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-04 20:38:40,171][riot01][DEBUG ] "min_mon_release": 0, -[2020-03-04 20:38:40,171][riot01][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-04 20:38:40,171][riot01][DEBUG ] "modified": "2020-03-04 20:34:06.060053", -[2020-03-04 20:38:40,171][riot01][DEBUG ] "mons": [ -[2020-03-04 20:38:40,171][riot01][DEBUG ] { -[2020-03-04 20:38:40,171][riot01][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-04 20:38:40,171][riot01][DEBUG ] "name": "pine01", -[2020-03-04 20:38:40,172][riot01][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-04 20:38:40,172][riot01][DEBUG ] "public_addrs": { -[2020-03-04 20:38:40,172][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:38:40,172][riot01][DEBUG ] { -[2020-03-04 20:38:40,172][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-04 20:38:40,172][riot01][DEBUG ] "nonce": 1, -[2020-03-04 20:38:40,172][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:38:40,173][riot01][DEBUG ] } -[2020-03-04 20:38:40,173][riot01][DEBUG ] ] -[2020-03-04 20:38:40,173][riot01][DEBUG ] }, -[2020-03-04 20:38:40,173][riot01][DEBUG ] "rank": 0 -[2020-03-04 20:38:40,173][riot01][DEBUG ] }, -[2020-03-04 20:38:40,173][riot01][DEBUG ] { -[2020-03-04 20:38:40,173][riot01][DEBUG ] "addr": "0.0.0.0:0/2", -[2020-03-04 20:38:40,174][riot01][DEBUG ] "name": "pine02", -[2020-03-04 20:38:40,174][riot01][DEBUG ] "public_addr": "0.0.0.0:0/2", -[2020-03-04 20:38:40,174][riot01][DEBUG ] "public_addrs": { -[2020-03-04 20:38:40,174][riot01][DEBUG ] "addrvec": [ -[2020-03-04 20:38:40,174][riot01][DEBUG ] { -[2020-03-04 20:38:40,174][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-04 20:38:40,175][riot01][DEBUG ] "nonce": 2, -[2020-03-04 20:38:40,175][riot01][DEBUG ] "type": "v1" -[2020-03-04 20:38:40,175][riot01][DEBUG ] } -[2020-03-04 20:38:40,175][riot01][DEBUG ] ] -[2020-03-04 20:38:40,175][riot01][DEBUG ] }, -[2020-03-04 20:38:40,175][riot01][DEBUG ] "rank": 1 -[2020-03-04 20:38:40,175][riot01][DEBUG ] } -[2020-03-04 20:38:40,176][riot01][DEBUG ] ] -[2020-03-04 20:38:40,176][riot01][DEBUG ] }, -[2020-03-04 20:38:40,176][riot01][DEBUG ] "name": "riot01", -[2020-03-04 20:38:40,176][riot01][DEBUG ] "outside_quorum": [], -[2020-03-04 20:38:40,176][riot01][DEBUG ] "quorum": [], -[2020-03-04 20:38:40,176][riot01][DEBUG ] "rank": -1, -[2020-03-04 20:38:40,176][riot01][DEBUG ] "state": "probing", -[2020-03-04 20:38:40,177][riot01][DEBUG ] "sync_provider": [] -[2020-03-04 20:38:40,177][riot01][DEBUG ] } -[2020-03-04 20:38:40,177][riot01][DEBUG ] ******************************************************************************** -[2020-03-04 20:38:40,177][riot01][INFO ] monitor: mon.riot01 is currently at the state of probing -[2020-03-04 20:39:51,523][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] username : root -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] verbose : False -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] quiet : False -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] func : -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-04 20:39:51,524][ceph_deploy.cli][INFO ] default_release : False -[2020-03-04 20:39:51,524][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-04 20:39:53,019][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-04 20:39:53,021][riot01][DEBUG ] detect platform information from remote host -[2020-03-04 20:39:53,190][riot01][DEBUG ] detect machine type -[2020-03-04 20:39:53,234][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:39:53,257][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-04 20:40:00,960][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-04 20:40:00,962][pine01][DEBUG ] detect platform information from remote host -[2020-03-04 20:40:01,437][pine01][DEBUG ] detect machine type -[2020-03-04 20:40:01,572][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:40:01,582][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-04 20:40:03,347][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-04 20:40:03,349][pine02][DEBUG ] detect platform information from remote host -[2020-03-04 20:40:03,506][pine02][DEBUG ] detect machine type -[2020-03-04 20:40:03,535][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:40:03,547][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-04 20:40:07,014][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-04 20:40:07,016][ebin01][DEBUG ] detect platform information from remote host -[2020-03-04 20:40:07,698][ebin01][DEBUG ] detect machine type -[2020-03-04 20:40:07,731][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:40:07,769][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-04 20:40:11,311][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-04 20:40:11,312][ebin02][DEBUG ] detect platform information from remote host -[2020-03-04 20:40:11,526][ebin02][DEBUG ] detect machine type -[2020-03-04 20:40:11,558][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-04 20:40:11,581][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-04 20:40:12,008][lenny][DEBUG ] connected to host: root@lenny -[2020-03-04 20:40:12,008][lenny][DEBUG ] detect platform information from remote host -[2020-03-04 20:40:12,023][lenny][DEBUG ] detect machine type -[2020-03-04 20:40:12,027][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:27:17,235][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add --address 192.168.10.164 riot01 -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] username : root -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] func : -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] address : 192.168.10.164 -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 19:27:17,236][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 19:27:17,237][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-06 19:27:17,237][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-06 19:27:18,740][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 19:27:18,751][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 19:27:18,918][riot01][DEBUG ] detect machine type -[2020-03-06 19:27:18,957][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:27:18,986][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-06 19:27:18,987][ceph_deploy.mon][DEBUG ] using mon address via --address 192.168.10.164 -[2020-03-06 19:27:18,987][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-06 19:27:20,549][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 19:27:20,551][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 19:27:20,719][riot01][DEBUG ] detect machine type -[2020-03-06 19:27:20,760][riot01][DEBUG ] find the location of an executable -[2020-03-06 19:27:20,767][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-06 19:27:20,768][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-06 19:27:20,768][riot01][DEBUG ] get remote short hostname -[2020-03-06 19:27:20,774][riot01][DEBUG ] adding mon to riot01 -[2020-03-06 19:27:20,775][riot01][DEBUG ] get remote short hostname -[2020-03-06 19:27:20,792][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:27:20,814][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-06 19:27:20,820][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-06 19:27:20,826][riot01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-06 19:27:20,833][riot01][DEBUG ] create the init path if it does not exist -[2020-03-06 19:27:20,849][riot01][INFO ] Running command: systemctl enable ceph.target -[2020-03-06 19:27:23,216][riot01][INFO ] Running command: systemctl enable ceph-mon@riot01 -[2020-03-06 19:27:25,466][riot01][INFO ] Running command: systemctl start ceph-mon@riot01 -[2020-03-06 19:27:27,610][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-06 19:27:29,705][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-06 19:27:30,577][riot01][DEBUG ] ******************************************************************************** -[2020-03-06 19:27:30,578][riot01][DEBUG ] status for monitor: mon.riot01 -[2020-03-06 19:27:30,579][riot01][DEBUG ] { -[2020-03-06 19:27:30,579][riot01][DEBUG ] "election_epoch": 0, -[2020-03-06 19:27:30,579][riot01][DEBUG ] "extra_probe_peers": [ -[2020-03-06 19:27:30,579][riot01][DEBUG ] { -[2020-03-06 19:27:30,579][riot01][DEBUG ] "addrvec": [ -[2020-03-06 19:27:30,579][riot01][DEBUG ] { -[2020-03-06 19:27:30,580][riot01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-06 19:27:30,580][riot01][DEBUG ] "nonce": 0, -[2020-03-06 19:27:30,580][riot01][DEBUG ] "type": "v2" -[2020-03-06 19:27:30,580][riot01][DEBUG ] }, -[2020-03-06 19:27:30,580][riot01][DEBUG ] { -[2020-03-06 19:27:30,580][riot01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-06 19:27:30,581][riot01][DEBUG ] "nonce": 0, -[2020-03-06 19:27:30,581][riot01][DEBUG ] "type": "v1" -[2020-03-06 19:27:30,581][riot01][DEBUG ] } -[2020-03-06 19:27:30,581][riot01][DEBUG ] ] -[2020-03-06 19:27:30,581][riot01][DEBUG ] }, -[2020-03-06 19:27:30,581][riot01][DEBUG ] { -[2020-03-06 19:27:30,581][riot01][DEBUG ] "addrvec": [ -[2020-03-06 19:27:30,581][riot01][DEBUG ] { -[2020-03-06 19:27:30,581][riot01][DEBUG ] "addr": "192.168.10.161:3300", -[2020-03-06 19:27:30,581][riot01][DEBUG ] "nonce": 0, -[2020-03-06 19:27:30,582][riot01][DEBUG ] "type": "v2" -[2020-03-06 19:27:30,582][riot01][DEBUG ] }, -[2020-03-06 19:27:30,582][riot01][DEBUG ] { -[2020-03-06 19:27:30,582][riot01][DEBUG ] "addr": "192.168.10.161:6789", -[2020-03-06 19:27:30,582][riot01][DEBUG ] "nonce": 0, -[2020-03-06 19:27:30,582][riot01][DEBUG ] "type": "v1" -[2020-03-06 19:27:30,582][riot01][DEBUG ] } -[2020-03-06 19:27:30,582][riot01][DEBUG ] ] -[2020-03-06 19:27:30,582][riot01][DEBUG ] } -[2020-03-06 19:27:30,582][riot01][DEBUG ] ], -[2020-03-06 19:27:30,582][riot01][DEBUG ] "feature_map": { -[2020-03-06 19:27:30,582][riot01][DEBUG ] "mon": [ -[2020-03-06 19:27:30,582][riot01][DEBUG ] { -[2020-03-06 19:27:30,582][riot01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-06 19:27:30,582][riot01][DEBUG ] "num": 1, -[2020-03-06 19:27:30,582][riot01][DEBUG ] "release": "luminous" -[2020-03-06 19:27:30,582][riot01][DEBUG ] } -[2020-03-06 19:27:30,582][riot01][DEBUG ] ] -[2020-03-06 19:27:30,582][riot01][DEBUG ] }, -[2020-03-06 19:27:30,582][riot01][DEBUG ] "features": { -[2020-03-06 19:27:30,583][riot01][DEBUG ] "quorum_con": "0", -[2020-03-06 19:27:30,583][riot01][DEBUG ] "quorum_mon": [], -[2020-03-06 19:27:30,583][riot01][DEBUG ] "required_con": "0", -[2020-03-06 19:27:30,583][riot01][DEBUG ] "required_mon": [] -[2020-03-06 19:27:30,583][riot01][DEBUG ] }, -[2020-03-06 19:27:30,583][riot01][DEBUG ] "monmap": { -[2020-03-06 19:27:30,583][riot01][DEBUG ] "created": "2020-03-04 20:34:06.060053", -[2020-03-06 19:27:30,583][riot01][DEBUG ] "epoch": 0, -[2020-03-06 19:27:30,583][riot01][DEBUG ] "features": { -[2020-03-06 19:27:30,583][riot01][DEBUG ] "optional": [], -[2020-03-06 19:27:30,583][riot01][DEBUG ] "persistent": [] -[2020-03-06 19:27:30,583][riot01][DEBUG ] }, -[2020-03-06 19:27:30,583][riot01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-06 19:27:30,583][riot01][DEBUG ] "min_mon_release": 0, -[2020-03-06 19:27:30,583][riot01][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-06 19:27:30,583][riot01][DEBUG ] "modified": "2020-03-04 20:34:06.060053", -[2020-03-06 19:27:30,583][riot01][DEBUG ] "mons": [ -[2020-03-06 19:27:30,583][riot01][DEBUG ] { -[2020-03-06 19:27:30,583][riot01][DEBUG ] "addr": ":/0", -[2020-03-06 19:27:30,583][riot01][DEBUG ] "name": "riot01", -[2020-03-06 19:27:30,584][riot01][DEBUG ] "public_addr": ":/0", -[2020-03-06 19:27:30,584][riot01][DEBUG ] "public_addrs": { -[2020-03-06 19:27:30,584][riot01][DEBUG ] "addrvec": [] -[2020-03-06 19:27:30,584][riot01][DEBUG ] }, -[2020-03-06 19:27:30,584][riot01][DEBUG ] "rank": 0 -[2020-03-06 19:27:30,584][riot01][DEBUG ] }, -[2020-03-06 19:27:30,584][riot01][DEBUG ] { -[2020-03-06 19:27:30,584][riot01][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-06 19:27:30,584][riot01][DEBUG ] "name": "pine01", -[2020-03-06 19:27:30,584][riot01][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-06 19:27:30,584][riot01][DEBUG ] "public_addrs": { -[2020-03-06 19:27:30,584][riot01][DEBUG ] "addrvec": [ -[2020-03-06 19:27:30,584][riot01][DEBUG ] { -[2020-03-06 19:27:30,584][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-06 19:27:30,584][riot01][DEBUG ] "nonce": 1, -[2020-03-06 19:27:30,584][riot01][DEBUG ] "type": "v1" -[2020-03-06 19:27:30,584][riot01][DEBUG ] } -[2020-03-06 19:27:30,584][riot01][DEBUG ] ] -[2020-03-06 19:27:30,584][riot01][DEBUG ] }, -[2020-03-06 19:27:30,584][riot01][DEBUG ] "rank": 1 -[2020-03-06 19:27:30,585][riot01][DEBUG ] }, -[2020-03-06 19:27:30,585][riot01][DEBUG ] { -[2020-03-06 19:27:30,585][riot01][DEBUG ] "addr": "0.0.0.0:0/2", -[2020-03-06 19:27:30,585][riot01][DEBUG ] "name": "pine02", -[2020-03-06 19:27:30,585][riot01][DEBUG ] "public_addr": "0.0.0.0:0/2", -[2020-03-06 19:27:30,585][riot01][DEBUG ] "public_addrs": { -[2020-03-06 19:27:30,585][riot01][DEBUG ] "addrvec": [ -[2020-03-06 19:27:30,585][riot01][DEBUG ] { -[2020-03-06 19:27:30,585][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-06 19:27:30,585][riot01][DEBUG ] "nonce": 2, -[2020-03-06 19:27:30,585][riot01][DEBUG ] "type": "v1" -[2020-03-06 19:27:30,585][riot01][DEBUG ] } -[2020-03-06 19:27:30,585][riot01][DEBUG ] ] -[2020-03-06 19:27:30,585][riot01][DEBUG ] }, -[2020-03-06 19:27:30,585][riot01][DEBUG ] "rank": 2 -[2020-03-06 19:27:30,585][riot01][DEBUG ] } -[2020-03-06 19:27:30,585][riot01][DEBUG ] ] -[2020-03-06 19:27:30,585][riot01][DEBUG ] }, -[2020-03-06 19:27:30,585][riot01][DEBUG ] "name": "riot01", -[2020-03-06 19:27:30,585][riot01][DEBUG ] "outside_quorum": [ -[2020-03-06 19:27:30,585][riot01][DEBUG ] "riot01" -[2020-03-06 19:27:30,586][riot01][DEBUG ] ], -[2020-03-06 19:27:30,586][riot01][DEBUG ] "quorum": [], -[2020-03-06 19:27:30,586][riot01][DEBUG ] "rank": -1, -[2020-03-06 19:27:30,586][riot01][DEBUG ] "state": "probing", -[2020-03-06 19:27:30,586][riot01][DEBUG ] "sync_provider": [] -[2020-03-06 19:27:30,586][riot01][DEBUG ] } -[2020-03-06 19:27:30,586][riot01][DEBUG ] ******************************************************************************** -[2020-03-06 19:27:30,586][riot01][INFO ] monitor: mon.riot01 is currently at the state of probing -[2020-03-06 19:32:18,606][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add --address 192.168.10.160 pine01 -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] username : root -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] mon : ['pine01'] -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] func : -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] address : 192.168.10.160 -[2020-03-06 19:32:18,606][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 19:32:18,607][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 19:32:18,607][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: pine01 -[2020-03-06 19:32:18,607][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to pine01 -[2020-03-06 19:32:27,036][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-06 19:32:27,038][pine01][DEBUG ] detect platform information from remote host -[2020-03-06 19:32:27,444][pine01][DEBUG ] detect machine type -[2020-03-06 19:32:27,468][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:32:27,901][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host pine01 -[2020-03-06 19:32:27,902][ceph_deploy.mon][DEBUG ] using mon address via --address 192.168.10.160 -[2020-03-06 19:32:27,902][ceph_deploy.mon][DEBUG ] detecting platform for host pine01 ... -[2020-03-06 19:32:34,520][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-06 19:32:34,522][pine01][DEBUG ] detect platform information from remote host -[2020-03-06 19:32:34,732][pine01][DEBUG ] detect machine type -[2020-03-06 19:32:34,967][pine01][DEBUG ] find the location of an executable -[2020-03-06 19:32:35,074][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-06 19:32:35,074][pine01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-06 19:32:35,074][pine01][DEBUG ] get remote short hostname -[2020-03-06 19:32:35,079][pine01][DEBUG ] adding mon to pine01 -[2020-03-06 19:32:35,080][pine01][DEBUG ] get remote short hostname -[2020-03-06 19:32:35,196][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:32:35,266][pine01][DEBUG ] create the mon path if it does not exist -[2020-03-06 19:32:35,375][pine01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-pine01/done -[2020-03-06 19:32:35,381][pine01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-06 19:32:35,443][pine01][DEBUG ] create the init path if it does not exist -[2020-03-06 19:32:35,511][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-06 19:32:36,661][pine01][INFO ] Running command: systemctl enable ceph-mon@pine01 -[2020-03-06 19:32:37,737][pine01][INFO ] Running command: systemctl start ceph-mon@pine01 -[2020-03-06 19:32:40,021][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-06 19:32:41,311][pine01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-06 19:32:42,386][pine01][DEBUG ] ******************************************************************************** -[2020-03-06 19:32:42,386][pine01][DEBUG ] status for monitor: mon.pine01 -[2020-03-06 19:32:42,387][pine01][DEBUG ] { -[2020-03-06 19:32:42,388][pine01][DEBUG ] "election_epoch": 92, -[2020-03-06 19:32:42,388][pine01][DEBUG ] "extra_probe_peers": [], -[2020-03-06 19:32:42,388][pine01][DEBUG ] "feature_map": { -[2020-03-06 19:32:42,388][pine01][DEBUG ] "client": [ -[2020-03-06 19:32:42,388][pine01][DEBUG ] { -[2020-03-06 19:32:42,389][pine01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-06 19:32:42,389][pine01][DEBUG ] "num": 1, -[2020-03-06 19:32:42,389][pine01][DEBUG ] "release": "luminous" -[2020-03-06 19:32:42,389][pine01][DEBUG ] } -[2020-03-06 19:32:42,389][pine01][DEBUG ] ], -[2020-03-06 19:32:42,389][pine01][DEBUG ] "mgr": [ -[2020-03-06 19:32:42,390][pine01][DEBUG ] { -[2020-03-06 19:32:42,390][pine01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-06 19:32:42,390][pine01][DEBUG ] "num": 1, -[2020-03-06 19:32:42,390][pine01][DEBUG ] "release": "luminous" -[2020-03-06 19:32:42,390][pine01][DEBUG ] } -[2020-03-06 19:32:42,390][pine01][DEBUG ] ], -[2020-03-06 19:32:42,391][pine01][DEBUG ] "mon": [ -[2020-03-06 19:32:42,391][pine01][DEBUG ] { -[2020-03-06 19:32:42,391][pine01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-06 19:32:42,391][pine01][DEBUG ] "num": 1, -[2020-03-06 19:32:42,391][pine01][DEBUG ] "release": "luminous" -[2020-03-06 19:32:42,391][pine01][DEBUG ] } -[2020-03-06 19:32:42,391][pine01][DEBUG ] ], -[2020-03-06 19:32:42,391][pine01][DEBUG ] "osd": [ -[2020-03-06 19:32:42,391][pine01][DEBUG ] { -[2020-03-06 19:32:42,391][pine01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-06 19:32:42,391][pine01][DEBUG ] "num": 4, -[2020-03-06 19:32:42,391][pine01][DEBUG ] "release": "luminous" -[2020-03-06 19:32:42,391][pine01][DEBUG ] } -[2020-03-06 19:32:42,392][pine01][DEBUG ] ] -[2020-03-06 19:32:42,392][pine01][DEBUG ] }, -[2020-03-06 19:32:42,392][pine01][DEBUG ] "features": { -[2020-03-06 19:32:42,392][pine01][DEBUG ] "quorum_con": "4611087854031667199", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "quorum_mon": [ -[2020-03-06 19:32:42,392][pine01][DEBUG ] "kraken", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "luminous", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "mimic", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "osdmap-prune", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "nautilus" -[2020-03-06 19:32:42,392][pine01][DEBUG ] ], -[2020-03-06 19:32:42,392][pine01][DEBUG ] "required_con": "2449958747315912708", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "required_mon": [ -[2020-03-06 19:32:42,392][pine01][DEBUG ] "kraken", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "luminous", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "mimic", -[2020-03-06 19:32:42,392][pine01][DEBUG ] "osdmap-prune", -[2020-03-06 19:32:42,393][pine01][DEBUG ] "nautilus" -[2020-03-06 19:32:42,393][pine01][DEBUG ] ] -[2020-03-06 19:32:42,393][pine01][DEBUG ] }, -[2020-03-06 19:32:42,393][pine01][DEBUG ] "monmap": { -[2020-03-06 19:32:42,393][pine01][DEBUG ] "created": "2020-03-03 20:33:27.003562", -[2020-03-06 19:32:42,393][pine01][DEBUG ] "epoch": 1, -[2020-03-06 19:32:42,393][pine01][DEBUG ] "features": { -[2020-03-06 19:32:42,393][pine01][DEBUG ] "optional": [], -[2020-03-06 19:32:42,393][pine01][DEBUG ] "persistent": [ -[2020-03-06 19:32:42,393][pine01][DEBUG ] "kraken", -[2020-03-06 19:32:42,393][pine01][DEBUG ] "luminous", -[2020-03-06 19:32:42,393][pine01][DEBUG ] "mimic", -[2020-03-06 19:32:42,393][pine01][DEBUG ] "osdmap-prune", -[2020-03-06 19:32:42,393][pine01][DEBUG ] "nautilus" -[2020-03-06 19:32:42,393][pine01][DEBUG ] ] -[2020-03-06 19:32:42,393][pine01][DEBUG ] }, -[2020-03-06 19:32:42,393][pine01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-06 19:32:42,393][pine01][DEBUG ] "min_mon_release": 14, -[2020-03-06 19:32:42,394][pine01][DEBUG ] "min_mon_release_name": "nautilus", -[2020-03-06 19:32:42,394][pine01][DEBUG ] "modified": "2020-03-03 20:33:27.003562", -[2020-03-06 19:32:42,394][pine01][DEBUG ] "mons": [ -[2020-03-06 19:32:42,394][pine01][DEBUG ] { -[2020-03-06 19:32:42,394][pine01][DEBUG ] "addr": "192.168.10.160:6789/0", -[2020-03-06 19:32:42,394][pine01][DEBUG ] "name": "pine01", -[2020-03-06 19:32:42,394][pine01][DEBUG ] "public_addr": "192.168.10.160:6789/0", -[2020-03-06 19:32:42,394][pine01][DEBUG ] "public_addrs": { -[2020-03-06 19:32:42,394][pine01][DEBUG ] "addrvec": [ -[2020-03-06 19:32:42,394][pine01][DEBUG ] { -[2020-03-06 19:32:42,394][pine01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-06 19:32:42,394][pine01][DEBUG ] "nonce": 0, -[2020-03-06 19:32:42,394][pine01][DEBUG ] "type": "v2" -[2020-03-06 19:32:42,394][pine01][DEBUG ] }, -[2020-03-06 19:32:42,394][pine01][DEBUG ] { -[2020-03-06 19:32:42,394][pine01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-06 19:32:42,394][pine01][DEBUG ] "nonce": 0, -[2020-03-06 19:32:42,395][pine01][DEBUG ] "type": "v1" -[2020-03-06 19:32:42,395][pine01][DEBUG ] } -[2020-03-06 19:32:42,395][pine01][DEBUG ] ] -[2020-03-06 19:32:42,395][pine01][DEBUG ] }, -[2020-03-06 19:32:42,395][pine01][DEBUG ] "rank": 0 -[2020-03-06 19:32:42,395][pine01][DEBUG ] }, -[2020-03-06 19:32:42,395][pine01][DEBUG ] { -[2020-03-06 19:32:42,395][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-06 19:32:42,395][pine01][DEBUG ] "name": "pine02", -[2020-03-06 19:32:42,395][pine01][DEBUG ] "public_addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-06 19:32:42,395][pine01][DEBUG ] "public_addrs": { -[2020-03-06 19:32:42,395][pine01][DEBUG ] "addrvec": [ -[2020-03-06 19:32:42,395][pine01][DEBUG ] { -[2020-03-06 19:32:42,395][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:3300", -[2020-03-06 19:32:42,395][pine01][DEBUG ] "nonce": 0, -[2020-03-06 19:32:42,395][pine01][DEBUG ] "type": "v2" -[2020-03-06 19:32:42,395][pine01][DEBUG ] }, -[2020-03-06 19:32:42,395][pine01][DEBUG ] { -[2020-03-06 19:32:42,396][pine01][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789", -[2020-03-06 19:32:42,396][pine01][DEBUG ] "nonce": 0, -[2020-03-06 19:32:42,396][pine01][DEBUG ] "type": "v1" -[2020-03-06 19:32:42,396][pine01][DEBUG ] } -[2020-03-06 19:32:42,396][pine01][DEBUG ] ] -[2020-03-06 19:32:42,396][pine01][DEBUG ] }, -[2020-03-06 19:32:42,396][pine01][DEBUG ] "rank": 1 -[2020-03-06 19:32:42,396][pine01][DEBUG ] } -[2020-03-06 19:32:42,396][pine01][DEBUG ] ] -[2020-03-06 19:32:42,396][pine01][DEBUG ] }, -[2020-03-06 19:32:42,396][pine01][DEBUG ] "name": "pine01", -[2020-03-06 19:32:42,396][pine01][DEBUG ] "outside_quorum": [], -[2020-03-06 19:32:42,396][pine01][DEBUG ] "quorum": [ -[2020-03-06 19:32:42,396][pine01][DEBUG ] 0, -[2020-03-06 19:32:42,396][pine01][DEBUG ] 1 -[2020-03-06 19:32:42,396][pine01][DEBUG ] ], -[2020-03-06 19:32:42,396][pine01][DEBUG ] "quorum_age": 1714, -[2020-03-06 19:32:42,397][pine01][DEBUG ] "rank": 0, -[2020-03-06 19:32:42,397][pine01][DEBUG ] "state": "leader", -[2020-03-06 19:32:42,397][pine01][DEBUG ] "sync_provider": [] -[2020-03-06 19:32:42,397][pine01][DEBUG ] } -[2020-03-06 19:32:42,397][pine01][DEBUG ] ******************************************************************************** -[2020-03-06 19:32:42,397][pine01][INFO ] monitor: mon.pine01 is running -[2020-03-06 19:32:57,323][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add --address 192.168.10.161 pine02 -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] username : root -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] mon : ['pine02'] -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] func : -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] address : 192.168.10.161 -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 19:32:57,324][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 19:32:57,325][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: pine02 -[2020-03-06 19:32:57,325][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to pine02 -[2020-03-06 19:32:59,113][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-06 19:32:59,115][pine02][DEBUG ] detect platform information from remote host -[2020-03-06 19:32:59,278][pine02][DEBUG ] detect machine type -[2020-03-06 19:32:59,305][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:32:59,326][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host pine02 -[2020-03-06 19:32:59,327][ceph_deploy.mon][DEBUG ] using mon address via --address 192.168.10.161 -[2020-03-06 19:32:59,327][ceph_deploy.mon][DEBUG ] detecting platform for host pine02 ... -[2020-03-06 19:33:00,798][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-06 19:33:00,800][pine02][DEBUG ] detect platform information from remote host -[2020-03-06 19:33:00,924][pine02][DEBUG ] detect machine type -[2020-03-06 19:33:00,952][pine02][DEBUG ] find the location of an executable -[2020-03-06 19:33:00,959][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-06 19:33:00,959][pine02][DEBUG ] determining if provided host has same hostname in remote -[2020-03-06 19:33:00,959][pine02][DEBUG ] get remote short hostname -[2020-03-06 19:33:00,966][pine02][DEBUG ] adding mon to pine02 -[2020-03-06 19:33:00,967][pine02][DEBUG ] get remote short hostname -[2020-03-06 19:33:00,987][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:33:01,003][pine02][DEBUG ] create the mon path if it does not exist -[2020-03-06 19:33:01,009][pine02][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-pine02/done -[2020-03-06 19:33:01,015][pine02][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-06 19:33:01,022][pine02][DEBUG ] create the init path if it does not exist -[2020-03-06 19:33:01,038][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-06 19:33:01,790][pine02][INFO ] Running command: systemctl enable ceph-mon@pine02 -[2020-03-06 19:33:02,525][pine02][INFO ] Running command: systemctl start ceph-mon@pine02 -[2020-03-06 19:33:04,608][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-06 19:33:05,542][pine02][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine02.asok mon_status -[2020-03-06 19:33:06,315][pine02][DEBUG ] ******************************************************************************** -[2020-03-06 19:33:06,315][pine02][DEBUG ] status for monitor: mon.pine02 -[2020-03-06 19:33:06,316][pine02][DEBUG ] { -[2020-03-06 19:33:06,317][pine02][DEBUG ] "election_epoch": 92, -[2020-03-06 19:33:06,317][pine02][DEBUG ] "extra_probe_peers": [], -[2020-03-06 19:33:06,317][pine02][DEBUG ] "feature_map": { -[2020-03-06 19:33:06,317][pine02][DEBUG ] "mgr": [ -[2020-03-06 19:33:06,317][pine02][DEBUG ] { -[2020-03-06 19:33:06,317][pine02][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-06 19:33:06,318][pine02][DEBUG ] "num": 1, -[2020-03-06 19:33:06,318][pine02][DEBUG ] "release": "luminous" -[2020-03-06 19:33:06,318][pine02][DEBUG ] } -[2020-03-06 19:33:06,318][pine02][DEBUG ] ], -[2020-03-06 19:33:06,318][pine02][DEBUG ] "mon": [ -[2020-03-06 19:33:06,318][pine02][DEBUG ] { -[2020-03-06 19:33:06,318][pine02][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-06 19:33:06,319][pine02][DEBUG ] "num": 1, -[2020-03-06 19:33:06,319][pine02][DEBUG ] "release": "luminous" -[2020-03-06 19:33:06,319][pine02][DEBUG ] } -[2020-03-06 19:33:06,319][pine02][DEBUG ] ] -[2020-03-06 19:33:06,319][pine02][DEBUG ] }, -[2020-03-06 19:33:06,319][pine02][DEBUG ] "features": { -[2020-03-06 19:33:06,319][pine02][DEBUG ] "quorum_con": "4611087854031667199", -[2020-03-06 19:33:06,319][pine02][DEBUG ] "quorum_mon": [ -[2020-03-06 19:33:06,319][pine02][DEBUG ] "kraken", -[2020-03-06 19:33:06,319][pine02][DEBUG ] "luminous", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "mimic", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "osdmap-prune", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "nautilus" -[2020-03-06 19:33:06,320][pine02][DEBUG ] ], -[2020-03-06 19:33:06,320][pine02][DEBUG ] "required_con": "2449958747315912708", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "required_mon": [ -[2020-03-06 19:33:06,320][pine02][DEBUG ] "kraken", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "luminous", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "mimic", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "osdmap-prune", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "nautilus" -[2020-03-06 19:33:06,320][pine02][DEBUG ] ] -[2020-03-06 19:33:06,320][pine02][DEBUG ] }, -[2020-03-06 19:33:06,320][pine02][DEBUG ] "monmap": { -[2020-03-06 19:33:06,320][pine02][DEBUG ] "created": "2020-03-03 20:33:27.003562", -[2020-03-06 19:33:06,320][pine02][DEBUG ] "epoch": 1, -[2020-03-06 19:33:06,320][pine02][DEBUG ] "features": { -[2020-03-06 19:33:06,320][pine02][DEBUG ] "optional": [], -[2020-03-06 19:33:06,320][pine02][DEBUG ] "persistent": [ -[2020-03-06 19:33:06,320][pine02][DEBUG ] "kraken", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "luminous", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "mimic", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "osdmap-prune", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "nautilus" -[2020-03-06 19:33:06,321][pine02][DEBUG ] ] -[2020-03-06 19:33:06,321][pine02][DEBUG ] }, -[2020-03-06 19:33:06,321][pine02][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "min_mon_release": 14, -[2020-03-06 19:33:06,321][pine02][DEBUG ] "min_mon_release_name": "nautilus", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "modified": "2020-03-03 20:33:27.003562", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "mons": [ -[2020-03-06 19:33:06,321][pine02][DEBUG ] { -[2020-03-06 19:33:06,321][pine02][DEBUG ] "addr": "192.168.10.160:6789/0", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "name": "pine01", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "public_addr": "192.168.10.160:6789/0", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "public_addrs": { -[2020-03-06 19:33:06,321][pine02][DEBUG ] "addrvec": [ -[2020-03-06 19:33:06,321][pine02][DEBUG ] { -[2020-03-06 19:33:06,321][pine02][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-06 19:33:06,321][pine02][DEBUG ] "nonce": 0, -[2020-03-06 19:33:06,322][pine02][DEBUG ] "type": "v2" -[2020-03-06 19:33:06,322][pine02][DEBUG ] }, -[2020-03-06 19:33:06,322][pine02][DEBUG ] { -[2020-03-06 19:33:06,322][pine02][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-06 19:33:06,322][pine02][DEBUG ] "nonce": 0, -[2020-03-06 19:33:06,322][pine02][DEBUG ] "type": "v1" -[2020-03-06 19:33:06,322][pine02][DEBUG ] } -[2020-03-06 19:33:06,322][pine02][DEBUG ] ] -[2020-03-06 19:33:06,322][pine02][DEBUG ] }, -[2020-03-06 19:33:06,322][pine02][DEBUG ] "rank": 0 -[2020-03-06 19:33:06,322][pine02][DEBUG ] }, -[2020-03-06 19:33:06,322][pine02][DEBUG ] { -[2020-03-06 19:33:06,322][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-06 19:33:06,322][pine02][DEBUG ] "name": "pine02", -[2020-03-06 19:33:06,322][pine02][DEBUG ] "public_addr": "[fd87:3937:d2b7::b5f]:6789/0", -[2020-03-06 19:33:06,322][pine02][DEBUG ] "public_addrs": { -[2020-03-06 19:33:06,322][pine02][DEBUG ] "addrvec": [ -[2020-03-06 19:33:06,322][pine02][DEBUG ] { -[2020-03-06 19:33:06,322][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:3300", -[2020-03-06 19:33:06,322][pine02][DEBUG ] "nonce": 0, -[2020-03-06 19:33:06,323][pine02][DEBUG ] "type": "v2" -[2020-03-06 19:33:06,323][pine02][DEBUG ] }, -[2020-03-06 19:33:06,323][pine02][DEBUG ] { -[2020-03-06 19:33:06,323][pine02][DEBUG ] "addr": "[fd87:3937:d2b7::b5f]:6789", -[2020-03-06 19:33:06,323][pine02][DEBUG ] "nonce": 0, -[2020-03-06 19:33:06,323][pine02][DEBUG ] "type": "v1" -[2020-03-06 19:33:06,323][pine02][DEBUG ] } -[2020-03-06 19:33:06,323][pine02][DEBUG ] ] -[2020-03-06 19:33:06,323][pine02][DEBUG ] }, -[2020-03-06 19:33:06,323][pine02][DEBUG ] "rank": 1 -[2020-03-06 19:33:06,323][pine02][DEBUG ] } -[2020-03-06 19:33:06,323][pine02][DEBUG ] ] -[2020-03-06 19:33:06,323][pine02][DEBUG ] }, -[2020-03-06 19:33:06,323][pine02][DEBUG ] "name": "pine02", -[2020-03-06 19:33:06,323][pine02][DEBUG ] "outside_quorum": [], -[2020-03-06 19:33:06,323][pine02][DEBUG ] "quorum": [ -[2020-03-06 19:33:06,323][pine02][DEBUG ] 0, -[2020-03-06 19:33:06,323][pine02][DEBUG ] 1 -[2020-03-06 19:33:06,323][pine02][DEBUG ] ], -[2020-03-06 19:33:06,323][pine02][DEBUG ] "quorum_age": 1738, -[2020-03-06 19:33:06,324][pine02][DEBUG ] "rank": 1, -[2020-03-06 19:33:06,324][pine02][DEBUG ] "state": "peon", -[2020-03-06 19:33:06,324][pine02][DEBUG ] "sync_provider": [] -[2020-03-06 19:33:06,324][pine02][DEBUG ] } -[2020-03-06 19:33:06,324][pine02][DEBUG ] ******************************************************************************** -[2020-03-06 19:33:06,324][pine02][INFO ] monitor: mon.pine02 is running -[2020-03-06 19:58:45,491][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] username : root -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] func : -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 19:58:45,492][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 19:58:45,492][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-06 19:58:47,023][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 19:58:47,025][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 19:58:47,193][riot01][DEBUG ] detect machine type -[2020-03-06 19:58:47,227][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:58:47,247][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-06 19:58:54,076][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-06 19:58:54,078][pine01][DEBUG ] detect platform information from remote host -[2020-03-06 19:58:54,245][pine01][DEBUG ] detect machine type -[2020-03-06 19:58:54,423][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:58:54,495][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-06 19:58:56,301][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-06 19:58:56,303][pine02][DEBUG ] detect platform information from remote host -[2020-03-06 19:58:56,451][pine02][DEBUG ] detect machine type -[2020-03-06 19:58:56,483][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:58:56,498][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-06 19:58:59,296][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-06 19:58:59,298][ebin01][DEBUG ] detect platform information from remote host -[2020-03-06 19:58:59,511][ebin01][DEBUG ] detect machine type -[2020-03-06 19:58:59,544][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:58:59,559][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-06 19:59:02,492][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-06 19:59:02,494][ebin02][DEBUG ] detect platform information from remote host -[2020-03-06 19:59:02,705][ebin02][DEBUG ] detect machine type -[2020-03-06 19:59:02,737][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 19:59:02,752][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-06 19:59:03,236][lenny][DEBUG ] connected to host: root@lenny -[2020-03-06 19:59:03,237][lenny][DEBUG ] detect platform information from remote host -[2020-03-06 19:59:03,252][lenny][DEBUG ] detect machine type -[2020-03-06 19:59:03,255][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 20:05:41,673][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 20:05:41,673][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purge riot01 -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] username : root -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] host : ['riot01'] -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] func : -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 20:05:41,683][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 20:05:41,683][ceph_deploy.install][INFO ] note that some dependencies *will not* be removed because they can cause issues with qemu-kvm -[2020-03-06 20:05:41,683][ceph_deploy.install][INFO ] like: librbd1 and librados2 -[2020-03-06 20:05:41,684][ceph_deploy.install][DEBUG ] Purging on cluster ceph hosts riot01 -[2020-03-06 20:05:41,684][ceph_deploy.install][DEBUG ] Detecting platform for host riot01 ... -[2020-03-06 20:05:43,218][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 20:05:43,220][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 20:05:43,393][riot01][DEBUG ] detect machine type -[2020-03-06 20:05:43,435][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-06 20:05:43,435][riot01][INFO ] Purging Ceph on riot01 -[2020-03-06 20:05:43,446][riot01][INFO ] Running command: env DEBIAN_FRONTEND=noninteractive DEBIAN_PRIORITY=critical apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw -[2020-03-06 20:05:43,975][riot01][DEBUG ] Reading package lists... -[2020-03-06 20:05:45,952][riot01][DEBUG ] Building dependency tree... -[2020-03-06 20:05:45,953][riot01][DEBUG ] Reading state information... -[2020-03-06 20:05:47,728][riot01][DEBUG ] Package 'ceph-fs-common' is not installed, so not removed -[2020-03-06 20:05:47,728][riot01][DEBUG ] The following packages were automatically installed and are no longer required: -[2020-03-06 20:05:47,729][riot01][DEBUG ] cryptsetup-bin dmeventd gdisk hdparm libaio1 libbabeltrace1 -[2020-03-06 20:05:47,729][riot01][DEBUG ] libboost-date-time1.67.0 libboost-program-options1.67.0 -[2020-03-06 20:05:47,730][riot01][DEBUG ] libboost-python1.67.0 libboost-random1.67.0 libboost-regex1.67.0 libcephfs2 -[2020-03-06 20:05:47,730][riot01][DEBUG ] libdevmapper-event1.02.1 libdw1 libfuse2 libgoogle-perftools4 libleveldb1d -[2020-03-06 20:05:47,730][riot01][DEBUG ] liblvm2cmd2.03 liboath0 libpython3.7 libradosstriper1 librbd1 libreadline5 -[2020-03-06 20:05:47,731][riot01][DEBUG ] libtcmalloc-minimal4 lvm2 python-pastedeploy-tpl python3-asn1crypto -[2020-03-06 20:05:47,731][riot01][DEBUG ] python3-bcrypt python3-bs4 python3-ceph-argparse python3-cephfs -[2020-03-06 20:05:47,731][riot01][DEBUG ] python3-certifi python3-cffi-backend python3-chardet python3-cherrypy3 -[2020-03-06 20:05:47,734][riot01][DEBUG ] python3-cryptography python3-idna python3-jwt python3-logutils python3-mako -[2020-03-06 20:05:47,738][riot01][DEBUG ] python3-markupsafe python3-openssl python3-paste python3-pastedeploy -[2020-03-06 20:05:47,738][riot01][DEBUG ] python3-pecan python3-pkg-resources python3-prettytable python3-rados -[2020-03-06 20:05:47,754][riot01][DEBUG ] python3-rbd python3-requests python3-simplegeneric python3-singledispatch -[2020-03-06 20:05:47,755][riot01][DEBUG ] python3-soupsieve python3-tempita python3-urllib3 python3-waitress -[2020-03-06 20:05:47,756][riot01][DEBUG ] python3-webob python3-webtest python3-werkzeug xfsprogs -[2020-03-06 20:05:47,760][riot01][DEBUG ] Use 'apt autoremove' to remove them. -[2020-03-06 20:05:48,380][riot01][DEBUG ] The following packages will be REMOVED: -[2020-03-06 20:05:48,380][riot01][DEBUG ] ceph* ceph-base* ceph-common* ceph-mds* ceph-mgr* ceph-mon* ceph-osd* -[2020-03-06 20:05:48,380][riot01][DEBUG ] radosgw* -[2020-03-06 20:05:52,422][riot01][DEBUG ] [master 1ebdf38] saving uncommitted changes in /etc prior to apt run -[2020-03-06 20:05:52,423][riot01][DEBUG ] 1 file changed, 13 insertions(+) -[2020-03-06 20:05:55,759][riot01][DEBUG ] 0 upgraded, 0 newly installed, 8 to remove and 3 not upgraded. -[2020-03-06 20:05:55,760][riot01][DEBUG ] After this operation, 168 MB disk space will be freed. -[2020-03-06 20:05:57,335][riot01][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 58434 files and directories currently installed.) -[2020-03-06 20:05:57,351][riot01][DEBUG ] Removing ceph-mds (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:03,261][riot01][DEBUG ] Removing ceph (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:03,578][riot01][DEBUG ] Removing ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:08,778][riot01][DEBUG ] Removing ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:14,834][riot01][DEBUG ] Removing radosgw (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:25,527][riot01][DEBUG ] Removing ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:32,180][riot01][DEBUG ] Removing ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:37,634][riot01][DEBUG ] Removing ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:43,641][riot01][DEBUG ] Processing triggers for libc-bin (2.28-10) ... -[2020-03-06 20:06:45,716][riot01][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 58012 files and directories currently installed.) -[2020-03-06 20:06:45,718][riot01][DEBUG ] Purging configuration files for radosgw (14.2.7-1~bpo10+1) ... -[2020-03-06 20:06:55,350][riot01][DEBUG ] Purging configuration files for ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-06 20:07:06,387][riot01][DEBUG ] dpkg: warning: while removing ceph-mon, directory '/var/lib/ceph/mon' not empty so not removed -[2020-03-06 20:07:06,502][riot01][DEBUG ] Purging configuration files for ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-06 20:07:13,817][riot01][DEBUG ] dpkg: warning: while removing ceph-base, directory '/var/lib/ceph/tmp' not empty so not removed -[2020-03-06 20:07:14,485][riot01][DEBUG ] Purging configuration files for ceph-mds (14.2.7-1~bpo10+1) ... -[2020-03-06 20:07:21,849][riot01][DEBUG ] Purging configuration files for ceph (14.2.7-1~bpo10+1) ... -[2020-03-06 20:07:22,166][riot01][DEBUG ] Purging configuration files for ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-06 20:07:30,229][riot01][DEBUG ] Purging configuration files for ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-06 20:07:40,663][riot01][DEBUG ] Purging configuration files for ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-06 20:07:48,628][riot01][DEBUG ] Processing triggers for systemd (241-7~deb10u3) ... -[2020-03-06 20:07:58,359][riot01][DEBUG ] [master be62a64] committing changes in /etc made by "apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw" -[2020-03-06 20:07:58,360][riot01][DEBUG ] 39 files changed, 2 insertions(+), 302 deletions(-) -[2020-03-06 20:07:58,361][riot01][DEBUG ] delete mode 100644 ceph/ceph.client.admin.keyring -[2020-03-06 20:07:58,361][riot01][DEBUG ] delete mode 100644 ceph/ceph.conf -[2020-03-06 20:07:58,362][riot01][DEBUG ] delete mode 100644 ceph/tmpSdszE4 -[2020-03-06 20:07:58,362][riot01][DEBUG ] delete mode 100644 default/ceph -[2020-03-06 20:07:58,362][riot01][DEBUG ] delete mode 100755 init.d/radosgw -[2020-03-06 20:07:58,362][riot01][DEBUG ] delete mode 100755 init.d/rbdmap -[2020-03-06 20:07:58,364][riot01][DEBUG ] delete mode 100644 logrotate.d/ceph-common -[2020-03-06 20:07:58,364][riot01][DEBUG ] delete mode 120000 rc0.d/K01radosgw -[2020-03-06 20:07:58,365][riot01][DEBUG ] delete mode 120000 rc0.d/K01rbdmap -[2020-03-06 20:07:58,369][riot01][DEBUG ] delete mode 120000 rc1.d/K01radosgw -[2020-03-06 20:07:58,370][riot01][DEBUG ] delete mode 120000 rc1.d/K01rbdmap -[2020-03-06 20:07:58,387][riot01][DEBUG ] delete mode 120000 rc2.d/S01radosgw -[2020-03-06 20:07:58,387][riot01][DEBUG ] delete mode 120000 rc2.d/S01rbdmap -[2020-03-06 20:07:58,391][riot01][DEBUG ] delete mode 120000 rc3.d/S01radosgw -[2020-03-06 20:07:58,393][riot01][DEBUG ] delete mode 120000 rc3.d/S01rbdmap -[2020-03-06 20:07:58,394][riot01][DEBUG ] delete mode 120000 rc4.d/S01radosgw -[2020-03-06 20:07:58,396][riot01][DEBUG ] delete mode 120000 rc4.d/S01rbdmap -[2020-03-06 20:07:58,399][riot01][DEBUG ] delete mode 120000 rc5.d/S01radosgw -[2020-03-06 20:07:58,400][riot01][DEBUG ] delete mode 120000 rc5.d/S01rbdmap -[2020-03-06 20:07:58,404][riot01][DEBUG ] delete mode 120000 rc6.d/K01radosgw -[2020-03-06 20:07:58,412][riot01][DEBUG ] delete mode 120000 rc6.d/K01rbdmap -[2020-03-06 20:07:58,414][riot01][DEBUG ] delete mode 100644 sudoers.d/ceph-osd-smartctl -[2020-03-06 20:07:58,417][riot01][DEBUG ] delete mode 100644 sysctl.d/30-ceph-osd.conf -[2020-03-06 20:07:58,425][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-crash.service -[2020-03-06 20:07:58,426][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mds.target -[2020-03-06 20:07:58,426][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mgr.target -[2020-03-06 20:07:58,434][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mon.target -[2020-03-06 20:07:58,438][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-osd.target -[2020-03-06 20:07:58,442][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-radosgw.target -[2020-03-06 20:07:58,450][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mds.target -[2020-03-06 20:07:58,450][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mgr.target -[2020-03-06 20:07:58,451][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mon.target -[2020-03-06 20:07:58,467][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-osd.target -[2020-03-06 20:07:58,471][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-radosgw.target -[2020-03-06 20:07:58,473][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph.target -[2020-03-06 20:07:58,474][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/rbdmap.service -[2020-03-06 20:07:58,640][riot01][WARNING] W: --force-yes is deprecated, use one of the options starting with --allow instead. -[2020-03-06 20:18:49,294][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purgedata riot01 -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] username : root -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] host : ['riot01'] -[2020-03-06 20:18:49,294][ceph_deploy.cli][INFO ] func : -[2020-03-06 20:18:49,295][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 20:18:49,295][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 20:18:49,295][ceph_deploy.install][DEBUG ] Purging data from cluster ceph hosts riot01 -[2020-03-06 20:18:50,774][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 20:18:50,776][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 20:18:50,946][riot01][DEBUG ] detect machine type -[2020-03-06 20:18:50,990][riot01][DEBUG ] find the location of an executable -[2020-03-06 20:18:52,623][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 20:18:52,625][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 20:18:52,897][riot01][DEBUG ] detect machine type -[2020-03-06 20:18:52,937][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-06 20:18:52,938][riot01][INFO ] purging data on riot01 -[2020-03-06 20:18:52,947][riot01][INFO ] Running command: rm -rf --one-file-system -- /var/lib/ceph -[2020-03-06 20:18:53,099][riot01][INFO ] Running command: rm -rf --one-file-system -- /etc/ceph/ -[2020-03-06 20:26:05,543][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] username : root -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] client : ['riot01'] -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] func : -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 20:26:05,544][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 20:26:05,544][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-06 20:26:07,026][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 20:26:07,027][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 20:26:07,194][riot01][DEBUG ] detect machine type -[2020-03-06 20:26:07,239][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 20:33:09,870][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] username : root -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] func : -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 20:33:09,871][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 20:33:09,871][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-06 20:33:11,310][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 20:33:11,312][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 20:33:11,483][riot01][DEBUG ] detect machine type -[2020-03-06 20:33:11,523][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 20:33:11,545][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-06 20:33:18,074][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-06 20:33:18,075][pine01][DEBUG ] detect platform information from remote host -[2020-03-06 20:33:18,345][pine01][DEBUG ] detect machine type -[2020-03-06 20:33:18,544][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 20:33:18,561][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-06 20:33:20,169][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-06 20:33:20,170][pine02][DEBUG ] detect platform information from remote host -[2020-03-06 20:33:20,294][pine02][DEBUG ] detect machine type -[2020-03-06 20:33:20,321][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 20:33:20,334][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-06 20:33:22,943][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-06 20:33:22,945][ebin01][DEBUG ] detect platform information from remote host -[2020-03-06 20:33:23,105][ebin01][DEBUG ] detect machine type -[2020-03-06 20:33:23,140][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 20:33:23,156][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-06 20:33:25,655][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-06 20:33:25,657][ebin02][DEBUG ] detect platform information from remote host -[2020-03-06 20:33:25,816][ebin02][DEBUG ] detect machine type -[2020-03-06 20:33:25,849][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 20:33:25,864][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-06 20:33:26,420][lenny][DEBUG ] connected to host: root@lenny -[2020-03-06 20:33:26,421][lenny][DEBUG ] detect platform information from remote host -[2020-03-06 20:33:26,435][lenny][DEBUG ] detect machine type -[2020-03-06 20:33:26,438][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:17:16,152][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 21:17:16,152][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-06 21:17:16,152][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 21:17:16,152][ceph_deploy.cli][INFO ] username : root -[2020-03-06 21:17:16,152][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 21:17:16,152][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-06 21:17:16,152][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-06 21:17:16,152][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 21:17:16,153][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 21:17:16,153][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 21:17:16,153][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-06 21:17:16,153][ceph_deploy.cli][INFO ] func : -[2020-03-06 21:17:16,153][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 21:17:16,153][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 21:17:16,153][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-06 21:17:17,642][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 21:17:17,643][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 21:17:17,812][riot01][DEBUG ] detect machine type -[2020-03-06 21:17:17,854][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:17:17,877][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-06 21:17:25,073][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-06 21:17:25,075][pine01][DEBUG ] detect platform information from remote host -[2020-03-06 21:17:25,358][pine01][DEBUG ] detect machine type -[2020-03-06 21:17:25,500][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:17:25,575][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-06 21:17:27,150][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-06 21:17:27,151][pine02][DEBUG ] detect platform information from remote host -[2020-03-06 21:17:27,277][pine02][DEBUG ] detect machine type -[2020-03-06 21:17:27,307][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:17:27,319][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-06 21:17:31,881][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-06 21:17:31,883][ebin01][DEBUG ] detect platform information from remote host -[2020-03-06 21:17:32,126][ebin01][DEBUG ] detect machine type -[2020-03-06 21:17:32,159][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:17:32,186][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-06 21:17:36,816][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-06 21:17:36,818][ebin02][DEBUG ] detect platform information from remote host -[2020-03-06 21:17:37,069][ebin02][DEBUG ] detect machine type -[2020-03-06 21:17:37,102][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:17:37,142][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-06 21:17:37,506][lenny][DEBUG ] connected to host: root@lenny -[2020-03-06 21:17:37,506][lenny][DEBUG ] detect platform information from remote host -[2020-03-06 21:17:37,521][lenny][DEBUG ] detect machine type -[2020-03-06 21:17:37,524][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:44:21,188][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] username : root -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] func : -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 21:44:21,188][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 21:44:21,188][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-06 21:44:22,649][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 21:44:22,651][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 21:44:22,821][riot01][DEBUG ] detect machine type -[2020-03-06 21:44:22,862][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:44:22,884][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-06 21:44:27,880][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-06 21:44:27,881][pine01][DEBUG ] detect platform information from remote host -[2020-03-06 21:44:28,122][pine01][DEBUG ] detect machine type -[2020-03-06 21:44:28,341][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:44:28,424][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-06 21:44:29,906][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-06 21:44:29,908][pine02][DEBUG ] detect platform information from remote host -[2020-03-06 21:44:30,031][pine02][DEBUG ] detect machine type -[2020-03-06 21:44:30,055][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:44:30,066][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-06 21:44:33,190][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-06 21:44:33,191][ebin01][DEBUG ] detect platform information from remote host -[2020-03-06 21:44:33,440][ebin01][DEBUG ] detect machine type -[2020-03-06 21:44:33,472][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:44:33,490][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-06 21:44:36,993][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-06 21:44:36,995][ebin02][DEBUG ] detect platform information from remote host -[2020-03-06 21:44:37,230][ebin02][DEBUG ] detect machine type -[2020-03-06 21:44:37,262][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:44:37,288][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-06 21:44:37,706][lenny][DEBUG ] connected to host: root@lenny -[2020-03-06 21:44:37,707][lenny][DEBUG ] detect platform information from remote host -[2020-03-06 21:44:37,722][lenny][DEBUG ] detect machine type -[2020-03-06 21:44:37,725][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:52:46,565][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add --address 192.168.10.164 riot01 -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] username : root -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] func : -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] address : 192.168.10.164 -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 21:52:46,565][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 21:52:46,566][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-06 21:52:46,566][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-06 21:52:48,049][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 21:52:48,051][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 21:52:48,219][riot01][DEBUG ] detect machine type -[2020-03-06 21:52:48,258][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:52:48,288][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-06 21:52:48,289][ceph_deploy.mon][DEBUG ] using mon address via --address 192.168.10.164 -[2020-03-06 21:52:48,289][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-06 21:52:49,888][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 21:52:49,890][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 21:52:50,061][riot01][DEBUG ] detect machine type -[2020-03-06 21:52:50,098][riot01][DEBUG ] find the location of an executable -[2020-03-06 21:52:50,104][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-06 21:52:50,105][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-06 21:52:50,105][riot01][DEBUG ] get remote short hostname -[2020-03-06 21:52:50,112][riot01][DEBUG ] adding mon to riot01 -[2020-03-06 21:52:50,112][riot01][DEBUG ] get remote short hostname -[2020-03-06 21:52:50,130][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:52:50,151][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-06 21:52:50,158][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-06 21:52:50,164][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-06 21:52:50,170][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-06 21:52:50,170][riot01][DEBUG ] create the monitor keyring file -[2020-03-06 21:52:50,187][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-06 21:52:50,971][riot01][WARNING] Traceback (most recent call last): -[2020-03-06 21:52:50,972][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-06 21:52:50,973][riot01][WARNING] retval = main() -[2020-03-06 21:52:50,973][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-06 21:52:50,973][riot01][WARNING] conffile=conffile) -[2020-03-06 21:52:50,974][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-06 21:52:50,974][riot01][WARNING] raise Exception("timed out") -[2020-03-06 21:52:50,974][riot01][WARNING] Exception: timed out -[2020-03-06 21:52:51,089][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-06 21:52:51,090][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-06 21:52:51,090][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-06 21:53:33,644][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-06 21:53:33,644][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add --address 192.168.10.164 riot01 -[2020-03-06 21:53:33,644][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-06 21:53:33,644][ceph_deploy.cli][INFO ] username : root -[2020-03-06 21:53:33,644][ceph_deploy.cli][INFO ] verbose : False -[2020-03-06 21:53:33,644][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-06 21:53:33,644][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-06 21:53:33,644][ceph_deploy.cli][INFO ] quiet : False -[2020-03-06 21:53:33,645][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-06 21:53:33,645][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-06 21:53:33,645][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-06 21:53:33,645][ceph_deploy.cli][INFO ] func : -[2020-03-06 21:53:33,645][ceph_deploy.cli][INFO ] address : 192.168.10.164 -[2020-03-06 21:53:33,645][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-06 21:53:33,645][ceph_deploy.cli][INFO ] default_release : False -[2020-03-06 21:53:33,645][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-06 21:53:33,645][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-06 21:53:35,205][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 21:53:35,207][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 21:53:35,381][riot01][DEBUG ] detect machine type -[2020-03-06 21:53:35,422][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:53:35,452][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-06 21:53:35,453][ceph_deploy.mon][DEBUG ] using mon address via --address 192.168.10.164 -[2020-03-06 21:53:35,453][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-06 21:53:36,984][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-06 21:53:36,985][riot01][DEBUG ] detect platform information from remote host -[2020-03-06 21:53:37,156][riot01][DEBUG ] detect machine type -[2020-03-06 21:53:37,198][riot01][DEBUG ] find the location of an executable -[2020-03-06 21:53:37,205][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-06 21:53:37,205][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-06 21:53:37,205][riot01][DEBUG ] get remote short hostname -[2020-03-06 21:53:37,212][riot01][DEBUG ] adding mon to riot01 -[2020-03-06 21:53:37,213][riot01][DEBUG ] get remote short hostname -[2020-03-06 21:53:37,230][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-06 21:53:37,252][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-06 21:53:37,257][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-06 21:53:37,264][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-06 21:53:37,270][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-06 21:53:37,271][riot01][DEBUG ] create the monitor keyring file -[2020-03-06 21:53:37,288][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-06 21:53:38,074][riot01][WARNING] Traceback (most recent call last): -[2020-03-06 21:53:38,074][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-06 21:53:38,074][riot01][WARNING] retval = main() -[2020-03-06 21:53:38,075][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-06 21:53:38,075][riot01][WARNING] conffile=conffile) -[2020-03-06 21:53:38,075][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-06 21:53:38,075][riot01][WARNING] raise Exception("timed out") -[2020-03-06 21:53:38,076][riot01][WARNING] Exception: timed out -[2020-03-06 21:53:38,191][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-06 21:53:38,191][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-06 21:53:38,191][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-07 17:04:27,480][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:04:27,481][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-07 17:04:27,481][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:04:27,481][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:04:27,481][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:04:27,481][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:04:27,482][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:04:27,482][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-07 17:04:29,776][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:04:29,789][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:04:29,952][riot01][DEBUG ] detect machine type -[2020-03-07 17:04:29,990][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:04:30,009][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-07 17:04:40,668][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 17:04:40,670][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 17:04:41,364][pine01][DEBUG ] detect machine type -[2020-03-07 17:04:41,506][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:04:41,570][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-07 17:04:43,344][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-07 17:04:43,346][pine02][DEBUG ] detect platform information from remote host -[2020-03-07 17:04:43,501][pine02][DEBUG ] detect machine type -[2020-03-07 17:04:43,528][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:04:43,561][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-07 17:04:46,073][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-07 17:04:46,075][ebin01][DEBUG ] detect platform information from remote host -[2020-03-07 17:04:46,290][ebin01][DEBUG ] detect machine type -[2020-03-07 17:04:46,323][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:04:46,337][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-07 17:04:48,769][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-07 17:04:48,770][ebin02][DEBUG ] detect platform information from remote host -[2020-03-07 17:04:48,983][ebin02][DEBUG ] detect machine type -[2020-03-07 17:04:49,012][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:04:49,026][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-07 17:04:49,442][lenny][DEBUG ] connected to host: root@lenny -[2020-03-07 17:04:49,443][lenny][DEBUG ] detect platform information from remote host -[2020-03-07 17:04:49,460][lenny][DEBUG ] detect machine type -[2020-03-07 17:04:49,464][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:13:42,396][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:13:42,396][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root gatherkeys pine01 -[2020-03-07 17:13:42,396][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:13:42,396][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:13:42,396][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:13:42,396][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:13:42,396][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:13:42,397][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:13:42,397][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:13:42,397][ceph_deploy.cli][INFO ] mon : ['pine01'] -[2020-03-07 17:13:42,397][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:13:42,397][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:13:42,397][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:13:42,407][ceph_deploy.gatherkeys][INFO ] Storing keys in temp directory /tmp/tmpgu3Qkm -[2020-03-07 17:13:49,132][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 17:13:49,134][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 17:13:49,344][pine01][DEBUG ] detect machine type -[2020-03-07 17:13:49,473][pine01][DEBUG ] get remote short hostname -[2020-03-07 17:13:49,528][pine01][DEBUG ] fetch remote file -[2020-03-07 17:13:49,548][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --admin-daemon=/var/run/ceph/ceph-mon.pine01.asok mon_status -[2020-03-07 17:13:50,805][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.admin -[2020-03-07 17:13:53,707][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-mds -[2020-03-07 17:13:56,582][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-mgr -[2020-03-07 17:13:59,303][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-osd -[2020-03-07 17:14:02,284][pine01][INFO ] Running command: /usr/bin/ceph --connect-timeout=25 --cluster=ceph --name mon. --keyring=/var/lib/ceph/mon/ceph-pine01/keyring auth get client.bootstrap-rgw -[2020-03-07 17:14:05,313][ceph_deploy.gatherkeys][INFO ] keyring 'ceph.client.admin.keyring' already exists -[2020-03-07 17:14:05,313][ceph_deploy.gatherkeys][INFO ] keyring 'ceph.bootstrap-mds.keyring' already exists -[2020-03-07 17:14:05,314][ceph_deploy.gatherkeys][INFO ] keyring 'ceph.bootstrap-mgr.keyring' already exists -[2020-03-07 17:14:05,315][ceph_deploy.gatherkeys][INFO ] keyring 'ceph.mon.keyring' already exists -[2020-03-07 17:14:05,315][ceph_deploy.gatherkeys][INFO ] keyring 'ceph.bootstrap-osd.keyring' already exists -[2020-03-07 17:14:05,316][ceph_deploy.gatherkeys][INFO ] keyring 'ceph.bootstrap-rgw.keyring' already exists -[2020-03-07 17:14:05,316][ceph_deploy.gatherkeys][INFO ] Destroy temp directory /tmp/tmpgu3Qkm -[2020-03-07 17:15:24,900][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create riot01 -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] rgw : [('riot01', 'rgw.riot01')] -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:15:24,900][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:15:24,901][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:15:24,901][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:15:24,901][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:15:24,901][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts riot01:rgw.riot01 -[2020-03-07 17:15:26,394][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:15:26,396][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:15:26,568][riot01][DEBUG ] detect machine type -[2020-03-07 17:15:26,611][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-07 17:15:26,612][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-07 17:15:26,614][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to riot01 -[2020-03-07 17:15:26,614][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:15:26,643][riot01][WARNING] rgw keyring does not exist yet, creating one -[2020-03-07 17:15:26,644][riot01][DEBUG ] create a keyring file -[2020-03-07 17:15:26,661][riot01][DEBUG ] create path recursively if it doesn't exist -[2020-03-07 17:15:26,679][riot01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.riot01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.riot01/keyring -[2020-03-07 17:15:29,069][riot01][ERROR ] Traceback (most recent call last): -[2020-03-07 17:15:29,069][riot01][ERROR ] File "/usr/bin/ceph", line 1266, in -[2020-03-07 17:15:29,070][riot01][ERROR ] retval = main() -[2020-03-07 17:15:29,070][riot01][ERROR ] File "/usr/bin/ceph", line 979, in main -[2020-03-07 17:15:29,071][riot01][ERROR ] conffile=conffile) -[2020-03-07 17:15:29,071][riot01][ERROR ] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-07 17:15:29,071][riot01][ERROR ] raise Exception("timed out") -[2020-03-07 17:15:29,072][riot01][ERROR ] Exception: timed out -[2020-03-07 17:15:29,072][riot01][ERROR ] exit code from command was: 1 -[2020-03-07 17:15:29,072][ceph_deploy.rgw][ERROR ] could not create rgw -[2020-03-07 17:15:29,072][ceph_deploy][ERROR ] GenericError: Failed to create 1 RGWs - -[2020-03-07 17:18:23,653][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purge riot01 -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] host : ['riot01'] -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:18:23,653][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:18:23,653][ceph_deploy.install][INFO ] note that some dependencies *will not* be removed because they can cause issues with qemu-kvm -[2020-03-07 17:18:23,653][ceph_deploy.install][INFO ] like: librbd1 and librados2 -[2020-03-07 17:18:23,653][ceph_deploy.install][DEBUG ] Purging on cluster ceph hosts riot01 -[2020-03-07 17:18:23,653][ceph_deploy.install][DEBUG ] Detecting platform for host riot01 ... -[2020-03-07 17:18:25,165][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:18:25,166][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:18:25,333][riot01][DEBUG ] detect machine type -[2020-03-07 17:18:25,372][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-07 17:18:25,373][riot01][INFO ] Purging Ceph on riot01 -[2020-03-07 17:18:25,383][riot01][INFO ] Running command: env DEBIAN_FRONTEND=noninteractive DEBIAN_PRIORITY=critical apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw -[2020-03-07 17:18:25,917][riot01][DEBUG ] Reading package lists... -[2020-03-07 17:18:29,750][riot01][DEBUG ] Building dependency tree... -[2020-03-07 17:18:29,750][riot01][DEBUG ] Reading state information... -[2020-03-07 17:18:31,524][riot01][DEBUG ] Package 'ceph-fs-common' is not installed, so not removed -[2020-03-07 17:18:31,524][riot01][DEBUG ] The following packages were automatically installed and are no longer required: -[2020-03-07 17:18:31,525][riot01][DEBUG ] cryptsetup-bin dmeventd gdisk hdparm libaio1 libbabeltrace1 -[2020-03-07 17:18:31,525][riot01][DEBUG ] libboost-date-time1.67.0 libboost-program-options1.67.0 -[2020-03-07 17:18:31,525][riot01][DEBUG ] libboost-python1.67.0 libboost-random1.67.0 libboost-regex1.67.0 libcephfs2 -[2020-03-07 17:18:31,526][riot01][DEBUG ] libdevmapper-event1.02.1 libdw1 libfuse2 libgoogle-perftools4 libleveldb1d -[2020-03-07 17:18:31,526][riot01][DEBUG ] liblvm2cmd2.03 liboath0 libpython3.7 libradosstriper1 librbd1 libreadline5 -[2020-03-07 17:18:31,530][riot01][DEBUG ] libtcmalloc-minimal4 lvm2 python-pastedeploy-tpl python3-asn1crypto -[2020-03-07 17:18:31,534][riot01][DEBUG ] python3-bcrypt python3-bs4 python3-ceph-argparse python3-cephfs -[2020-03-07 17:18:31,538][riot01][DEBUG ] python3-certifi python3-cffi-backend python3-chardet python3-cherrypy3 -[2020-03-07 17:18:31,540][riot01][DEBUG ] python3-cryptography python3-idna python3-jwt python3-logutils python3-mako -[2020-03-07 17:18:31,548][riot01][DEBUG ] python3-markupsafe python3-openssl python3-paste python3-pastedeploy -[2020-03-07 17:18:31,548][riot01][DEBUG ] python3-pecan python3-pkg-resources python3-prettytable python3-rados -[2020-03-07 17:18:31,550][riot01][DEBUG ] python3-rbd python3-requests python3-simplegeneric python3-singledispatch -[2020-03-07 17:18:31,552][riot01][DEBUG ] python3-soupsieve python3-tempita python3-urllib3 python3-waitress -[2020-03-07 17:18:31,568][riot01][DEBUG ] python3-webob python3-webtest python3-werkzeug xfsprogs -[2020-03-07 17:18:31,569][riot01][DEBUG ] Use 'apt autoremove' to remove them. -[2020-03-07 17:18:32,135][riot01][DEBUG ] The following packages will be REMOVED: -[2020-03-07 17:18:32,140][riot01][DEBUG ] ceph* ceph-base* ceph-common* ceph-mds* ceph-mgr* ceph-mon* ceph-osd* -[2020-03-07 17:18:32,144][riot01][DEBUG ] radosgw* -[2020-03-07 17:18:36,230][riot01][DEBUG ] [master 7f28d9f] saving uncommitted changes in /etc prior to apt run -[2020-03-07 17:18:36,230][riot01][DEBUG ] 3 files changed, 29 insertions(+) -[2020-03-07 17:18:36,230][riot01][DEBUG ] create mode 100644 ceph/ceph.conf -[2020-03-07 17:18:36,230][riot01][DEBUG ] create mode 100644 ceph/tmpMN9TB6 -[2020-03-07 17:18:38,305][riot01][DEBUG ] 0 upgraded, 0 newly installed, 8 to remove and 3 not upgraded. -[2020-03-07 17:18:38,305][riot01][DEBUG ] After this operation, 168 MB disk space will be freed. -[2020-03-07 17:18:38,671][riot01][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 58434 files and directories currently installed.) -[2020-03-07 17:18:38,688][riot01][DEBUG ] Removing ceph-mds (14.2.7-1~bpo10+1) ... -[2020-03-07 17:18:43,829][riot01][DEBUG ] Removing ceph (14.2.7-1~bpo10+1) ... -[2020-03-07 17:18:44,194][riot01][DEBUG ] Removing ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-07 17:18:49,386][riot01][DEBUG ] Removing ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-07 17:18:55,921][riot01][DEBUG ] Removing radosgw (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:03,967][riot01][DEBUG ] Removing ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:09,251][riot01][DEBUG ] Removing ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:14,535][riot01][DEBUG ] Removing ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:20,521][riot01][DEBUG ] Processing triggers for libc-bin (2.28-10) ... -[2020-03-07 17:19:23,245][riot01][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 58012 files and directories currently installed.) -[2020-03-07 17:19:23,261][riot01][DEBUG ] Purging configuration files for radosgw (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:34,718][riot01][DEBUG ] dpkg: warning: while removing radosgw, directory '/var/lib/ceph/radosgw' not empty so not removed -[2020-03-07 17:19:34,833][riot01][DEBUG ] Purging configuration files for ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:43,277][riot01][DEBUG ] Purging configuration files for ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:50,612][riot01][DEBUG ] dpkg: warning: while removing ceph-base, directory '/var/lib/ceph/bootstrap-rgw' not empty so not removed -[2020-03-07 17:19:50,726][riot01][DEBUG ] Purging configuration files for ceph-mds (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:58,067][riot01][DEBUG ] Purging configuration files for ceph (14.2.7-1~bpo10+1) ... -[2020-03-07 17:19:58,382][riot01][DEBUG ] Purging configuration files for ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-07 17:20:05,673][riot01][DEBUG ] Purging configuration files for ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-07 17:20:15,424][riot01][DEBUG ] Purging configuration files for ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-07 17:20:23,521][riot01][DEBUG ] Processing triggers for systemd (241-7~deb10u3) ... -[2020-03-07 17:20:33,323][riot01][DEBUG ] [master 0e8f3d6] committing changes in /etc made by "apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw" -[2020-03-07 17:20:33,323][riot01][DEBUG ] 40 files changed, 2 insertions(+), 300 deletions(-) -[2020-03-07 17:20:33,324][riot01][DEBUG ] delete mode 100644 ceph/ceph.conf -[2020-03-07 17:20:33,327][riot01][DEBUG ] delete mode 100644 ceph/rbdmap -[2020-03-07 17:20:33,336][riot01][DEBUG ] delete mode 100644 ceph/tmpMN9TB6 -[2020-03-07 17:20:33,337][riot01][DEBUG ] delete mode 100644 ceph/tmpg_aayV -[2020-03-07 17:20:33,339][riot01][DEBUG ] delete mode 100644 default/ceph -[2020-03-07 17:20:33,341][riot01][DEBUG ] delete mode 100755 init.d/radosgw -[2020-03-07 17:20:33,345][riot01][DEBUG ] delete mode 100755 init.d/rbdmap -[2020-03-07 17:20:33,345][riot01][DEBUG ] delete mode 100644 logrotate.d/ceph-common -[2020-03-07 17:20:33,347][riot01][DEBUG ] delete mode 120000 rc0.d/K01radosgw -[2020-03-07 17:20:33,351][riot01][DEBUG ] delete mode 120000 rc0.d/K01rbdmap -[2020-03-07 17:20:33,353][riot01][DEBUG ] delete mode 120000 rc1.d/K01radosgw -[2020-03-07 17:20:33,369][riot01][DEBUG ] delete mode 120000 rc1.d/K01rbdmap -[2020-03-07 17:20:33,369][riot01][DEBUG ] delete mode 120000 rc2.d/S01radosgw -[2020-03-07 17:20:33,370][riot01][DEBUG ] delete mode 120000 rc2.d/S01rbdmap -[2020-03-07 17:20:33,378][riot01][DEBUG ] delete mode 120000 rc3.d/S01radosgw -[2020-03-07 17:20:33,378][riot01][DEBUG ] delete mode 120000 rc3.d/S01rbdmap -[2020-03-07 17:20:33,379][riot01][DEBUG ] delete mode 120000 rc4.d/S01radosgw -[2020-03-07 17:20:33,383][riot01][DEBUG ] delete mode 120000 rc4.d/S01rbdmap -[2020-03-07 17:20:33,384][riot01][DEBUG ] delete mode 120000 rc5.d/S01radosgw -[2020-03-07 17:20:33,385][riot01][DEBUG ] delete mode 120000 rc5.d/S01rbdmap -[2020-03-07 17:20:33,387][riot01][DEBUG ] delete mode 120000 rc6.d/K01radosgw -[2020-03-07 17:20:33,403][riot01][DEBUG ] delete mode 120000 rc6.d/K01rbdmap -[2020-03-07 17:20:33,404][riot01][DEBUG ] delete mode 100644 sudoers.d/ceph-osd-smartctl -[2020-03-07 17:20:33,404][riot01][DEBUG ] delete mode 100644 sysctl.d/30-ceph-osd.conf -[2020-03-07 17:20:33,412][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-crash.service -[2020-03-07 17:20:33,413][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mds.target -[2020-03-07 17:20:33,416][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mgr.target -[2020-03-07 17:20:33,417][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mon.target -[2020-03-07 17:20:33,433][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-osd.target -[2020-03-07 17:20:33,434][riot01][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-radosgw.target -[2020-03-07 17:20:33,435][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mds.target -[2020-03-07 17:20:33,435][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mgr.target -[2020-03-07 17:20:33,435][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mon.target -[2020-03-07 17:20:33,439][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-osd.target -[2020-03-07 17:20:33,441][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-radosgw.target -[2020-03-07 17:20:33,457][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph.target -[2020-03-07 17:20:33,458][riot01][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/rbdmap.service -[2020-03-07 17:20:33,672][riot01][WARNING] W: --force-yes is deprecated, use one of the options starting with --allow instead. -[2020-03-07 17:20:48,971][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purgedata riot01 -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] host : ['riot01'] -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:20:48,971][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:20:48,971][ceph_deploy.install][DEBUG ] Purging data from cluster ceph hosts riot01 -[2020-03-07 17:20:50,467][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:20:50,470][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:20:50,638][riot01][DEBUG ] detect machine type -[2020-03-07 17:20:50,677][riot01][DEBUG ] find the location of an executable -[2020-03-07 17:20:52,270][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:20:52,272][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:20:52,441][riot01][DEBUG ] detect machine type -[2020-03-07 17:20:52,478][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-07 17:20:52,478][riot01][INFO ] purging data on riot01 -[2020-03-07 17:20:52,488][riot01][INFO ] Running command: rm -rf --one-file-system -- /var/lib/ceph -[2020-03-07 17:20:52,582][riot01][INFO ] Running command: rm -rf --one-file-system -- /etc/ceph/ -[2020-03-07 17:37:15,905][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy mon add riot01 -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] username : None -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] address : None -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:37:15,906][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:37:15,907][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-07 17:37:15,907][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-07 17:37:17,836][ceph_deploy][ERROR ] KeyboardInterrupt - -[2020-03-07 17:37:39,493][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:37:39,493][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:37:39,494][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:37:39,494][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-07 17:37:39,494][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:37:39,494][ceph_deploy.cli][INFO ] address : None -[2020-03-07 17:37:39,494][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:37:39,494][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:37:39,494][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-07 17:37:39,494][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-07 17:37:41,021][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:37:41,024][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:37:41,197][riot01][DEBUG ] detect machine type -[2020-03-07 17:37:41,238][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:37:41,264][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-07 17:37:41,269][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-07 17:37:41,270][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-07 17:37:42,858][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:37:42,860][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:37:43,029][riot01][DEBUG ] detect machine type -[2020-03-07 17:37:43,071][riot01][DEBUG ] find the location of an executable -[2020-03-07 17:37:43,078][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-07 17:37:43,078][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-07 17:37:43,079][riot01][DEBUG ] get remote short hostname -[2020-03-07 17:37:43,085][riot01][DEBUG ] adding mon to riot01 -[2020-03-07 17:37:43,085][riot01][DEBUG ] get remote short hostname -[2020-03-07 17:37:43,103][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:37:43,124][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-07 17:37:43,131][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 17:37:43,137][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 17:37:43,143][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-07 17:37:43,143][riot01][DEBUG ] create the monitor keyring file -[2020-03-07 17:37:43,161][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-07 17:37:43,949][riot01][WARNING] Traceback (most recent call last): -[2020-03-07 17:37:43,950][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-07 17:37:43,950][riot01][WARNING] retval = main() -[2020-03-07 17:37:43,951][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-07 17:37:43,951][riot01][WARNING] conffile=conffile) -[2020-03-07 17:37:43,951][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-07 17:37:43,952][riot01][WARNING] raise Exception("timed out") -[2020-03-07 17:37:43,952][riot01][WARNING] Exception: timed out -[2020-03-07 17:37:44,077][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-07 17:37:44,077][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-07 17:37:44,078][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-07 17:43:31,174][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:43:31,174][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin riot01 -[2020-03-07 17:43:31,174][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] client : ['riot01'] -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:43:31,175][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:43:31,175][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-07 17:43:32,688][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:43:32,690][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:43:32,859][riot01][DEBUG ] detect machine type -[2020-03-07 17:43:32,897][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:43:42,661][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon add riot01 -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] subcommand : add -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] address : None -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:43:42,662][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:43:42,663][ceph_deploy.mon][INFO ] ensuring configuration of new mon host: riot01 -[2020-03-07 17:43:42,663][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-07 17:43:44,136][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:43:44,138][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:43:44,313][riot01][DEBUG ] detect machine type -[2020-03-07 17:43:44,355][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:43:44,387][ceph_deploy.mon][DEBUG ] Adding mon to cluster ceph, host riot01 -[2020-03-07 17:43:44,392][ceph_deploy.mon][DEBUG ] using mon address by resolving host: 192.168.10.164 -[2020-03-07 17:43:44,392][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-07 17:43:45,976][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:43:45,978][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:43:46,146][riot01][DEBUG ] detect machine type -[2020-03-07 17:43:46,189][riot01][DEBUG ] find the location of an executable -[2020-03-07 17:43:46,195][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-07 17:43:46,196][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-07 17:43:46,196][riot01][DEBUG ] get remote short hostname -[2020-03-07 17:43:46,201][riot01][DEBUG ] adding mon to riot01 -[2020-03-07 17:43:46,202][riot01][DEBUG ] get remote short hostname -[2020-03-07 17:43:46,220][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:43:46,242][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-07 17:43:46,248][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 17:43:46,255][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 17:43:46,261][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-07 17:43:46,261][riot01][DEBUG ] create the monitor keyring file -[2020-03-07 17:43:46,279][riot01][INFO ] Running command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-07 17:43:47,068][riot01][WARNING] Traceback (most recent call last): -[2020-03-07 17:43:47,069][riot01][WARNING] File "/usr/bin/ceph", line 1266, in -[2020-03-07 17:43:47,069][riot01][WARNING] retval = main() -[2020-03-07 17:43:47,070][riot01][WARNING] File "/usr/bin/ceph", line 979, in main -[2020-03-07 17:43:47,070][riot01][WARNING] conffile=conffile) -[2020-03-07 17:43:47,071][riot01][WARNING] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-07 17:43:47,071][riot01][WARNING] raise Exception("timed out") -[2020-03-07 17:43:47,071][riot01][WARNING] Exception: timed out -[2020-03-07 17:43:47,187][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-07 17:43:47,187][ceph_deploy.mon][ERROR ] Failed to execute command: ceph --cluster ceph mon getmap -o /var/lib/ceph/tmp/ceph.riot01.monmap -[2020-03-07 17:43:47,188][ceph_deploy][ERROR ] GenericError: Failed to add monitor to host: riot01 - -[2020-03-07 17:56:25,332][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 17:56:25,332][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-07 17:56:25,332][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 17:56:25,332][ceph_deploy.cli][INFO ] username : root -[2020-03-07 17:56:25,332][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 17:56:25,332][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 17:56:25,332][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-07 17:56:25,332][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 17:56:25,333][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 17:56:25,333][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 17:56:25,333][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-07 17:56:25,333][ceph_deploy.cli][INFO ] func : -[2020-03-07 17:56:25,333][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 17:56:25,333][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 17:56:25,333][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-07 17:56:26,830][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 17:56:26,833][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 17:56:27,002][riot01][DEBUG ] detect machine type -[2020-03-07 17:56:27,094][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:56:27,125][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-07 17:56:34,115][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 17:56:34,118][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 17:56:34,319][pine01][DEBUG ] detect machine type -[2020-03-07 17:56:34,450][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:56:34,505][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-07 17:56:36,353][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-07 17:56:36,355][pine02][DEBUG ] detect platform information from remote host -[2020-03-07 17:56:36,475][pine02][DEBUG ] detect machine type -[2020-03-07 17:56:36,502][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:56:36,515][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-07 17:56:39,032][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-07 17:56:39,032][ebin01][DEBUG ] detect platform information from remote host -[2020-03-07 17:56:39,244][ebin01][DEBUG ] detect machine type -[2020-03-07 17:56:39,274][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:56:39,300][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-07 17:56:42,242][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-07 17:56:42,243][ebin02][DEBUG ] detect platform information from remote host -[2020-03-07 17:56:42,452][ebin02][DEBUG ] detect machine type -[2020-03-07 17:56:42,486][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 17:56:42,509][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-07 17:56:43,078][lenny][DEBUG ] connected to host: root@lenny -[2020-03-07 17:56:43,079][lenny][DEBUG ] detect platform information from remote host -[2020-03-07 17:56:43,103][lenny][DEBUG ] detect machine type -[2020-03-07 17:56:43,107][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:00:25,262][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 18:00:25,262][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf mon create riot01 -[2020-03-07 18:00:25,272][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 18:00:25,272][ceph_deploy.cli][INFO ] username : root -[2020-03-07 18:00:25,273][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 18:00:25,273][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 18:00:25,273][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 18:00:25,273][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 18:00:25,273][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 18:00:25,273][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 18:00:25,273][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-07 18:00:25,274][ceph_deploy.cli][INFO ] func : -[2020-03-07 18:00:25,274][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 18:00:25,274][ceph_deploy.cli][INFO ] keyrings : None -[2020-03-07 18:00:25,274][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 18:00:25,275][ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts riot01 -[2020-03-07 18:00:25,275][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-07 18:00:26,798][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 18:00:26,800][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 18:00:26,968][riot01][DEBUG ] detect machine type -[2020-03-07 18:00:27,055][riot01][DEBUG ] find the location of an executable -[2020-03-07 18:00:27,061][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-07 18:00:27,062][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-07 18:00:27,062][riot01][DEBUG ] get remote short hostname -[2020-03-07 18:00:27,068][riot01][DEBUG ] deploying mon to riot01 -[2020-03-07 18:00:27,069][riot01][DEBUG ] get remote short hostname -[2020-03-07 18:00:27,074][riot01][DEBUG ] remote hostname: riot01 -[2020-03-07 18:00:27,085][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:00:27,124][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-07 18:00:27,133][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 18:00:27,146][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 18:00:27,153][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-07 18:00:27,153][riot01][DEBUG ] create the monitor keyring file -[2020-03-07 18:00:27,194][riot01][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i riot01 --keyring /var/lib/ceph/tmp/ceph-riot01.mon.keyring --setuser 64045 --setgroup 64045 -[2020-03-07 18:00:28,382][riot01][WARNING] 2020-03-07 18:00:28.321 afb202c0 -1 ceph-mon: error opening mon data directory at '/var/lib/ceph/mon/ceph-riot01': (13) Permission denied -[2020-03-07 18:00:28,448][riot01][ERROR ] RuntimeError: command returned non-zero exit status: 1 -[2020-03-07 18:00:28,448][ceph_deploy.mon][ERROR ] Failed to execute command: ceph-mon --cluster ceph --mkfs -i riot01 --keyring /var/lib/ceph/tmp/ceph-riot01.mon.keyring --setuser 64045 --setgroup 64045 -[2020-03-07 18:00:28,449][ceph_deploy][ERROR ] GenericError: Failed to create 1 monitors - -[2020-03-07 18:01:17,355][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 18:01:17,355][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf mon create riot01 -[2020-03-07 18:01:17,355][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 18:01:17,355][ceph_deploy.cli][INFO ] username : root -[2020-03-07 18:01:17,355][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 18:01:17,355][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] func : -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] keyrings : None -[2020-03-07 18:01:17,356][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 18:01:17,356][ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts riot01 -[2020-03-07 18:01:17,356][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-07 18:01:18,870][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 18:01:18,872][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 18:01:19,041][riot01][DEBUG ] detect machine type -[2020-03-07 18:01:19,083][riot01][DEBUG ] find the location of an executable -[2020-03-07 18:01:19,089][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-07 18:01:19,090][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-07 18:01:19,090][riot01][DEBUG ] get remote short hostname -[2020-03-07 18:01:19,096][riot01][DEBUG ] deploying mon to riot01 -[2020-03-07 18:01:19,096][riot01][DEBUG ] get remote short hostname -[2020-03-07 18:01:19,103][riot01][DEBUG ] remote hostname: riot01 -[2020-03-07 18:01:19,115][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:01:19,137][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-07 18:01:19,143][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 18:01:19,150][riot01][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 18:01:19,157][riot01][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-07 18:01:19,157][riot01][DEBUG ] create the monitor keyring file -[2020-03-07 18:01:19,175][riot01][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i riot01 --keyring /var/lib/ceph/tmp/ceph-riot01.mon.keyring --setuser 64045 --setgroup 64045 -[2020-03-07 18:01:19,913][riot01][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-riot01.mon.keyring -[2020-03-07 18:01:19,920][riot01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-07 18:01:19,927][riot01][DEBUG ] create the init path if it does not exist -[2020-03-07 18:01:19,944][riot01][INFO ] Running command: systemctl enable ceph.target -[2020-03-07 18:01:22,192][riot01][INFO ] Running command: systemctl enable ceph-mon@riot01 -[2020-03-07 18:01:24,435][riot01][INFO ] Running command: systemctl start ceph-mon@riot01 -[2020-03-07 18:01:26,625][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-07 18:01:29,002][riot01][DEBUG ] ******************************************************************************** -[2020-03-07 18:01:29,003][riot01][DEBUG ] status for monitor: mon.riot01 -[2020-03-07 18:01:29,004][riot01][DEBUG ] { -[2020-03-07 18:01:29,004][riot01][DEBUG ] "election_epoch": 0, -[2020-03-07 18:01:29,004][riot01][DEBUG ] "extra_probe_peers": [ -[2020-03-07 18:01:29,004][riot01][DEBUG ] { -[2020-03-07 18:01:29,005][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:01:29,005][riot01][DEBUG ] { -[2020-03-07 18:01:29,005][riot01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-07 18:01:29,005][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:01:29,005][riot01][DEBUG ] "type": "v2" -[2020-03-07 18:01:29,005][riot01][DEBUG ] }, -[2020-03-07 18:01:29,006][riot01][DEBUG ] { -[2020-03-07 18:01:29,006][riot01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-07 18:01:29,006][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:01:29,006][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:01:29,007][riot01][DEBUG ] } -[2020-03-07 18:01:29,007][riot01][DEBUG ] ] -[2020-03-07 18:01:29,007][riot01][DEBUG ] }, -[2020-03-07 18:01:29,007][riot01][DEBUG ] { -[2020-03-07 18:01:29,007][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:01:29,007][riot01][DEBUG ] { -[2020-03-07 18:01:29,008][riot01][DEBUG ] "addr": "192.168.10.161:3300", -[2020-03-07 18:01:29,008][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:01:29,008][riot01][DEBUG ] "type": "v2" -[2020-03-07 18:01:29,008][riot01][DEBUG ] }, -[2020-03-07 18:01:29,008][riot01][DEBUG ] { -[2020-03-07 18:01:29,008][riot01][DEBUG ] "addr": "192.168.10.161:6789", -[2020-03-07 18:01:29,009][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:01:29,009][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:01:29,009][riot01][DEBUG ] } -[2020-03-07 18:01:29,009][riot01][DEBUG ] ] -[2020-03-07 18:01:29,009][riot01][DEBUG ] } -[2020-03-07 18:01:29,010][riot01][DEBUG ] ], -[2020-03-07 18:01:29,010][riot01][DEBUG ] "feature_map": { -[2020-03-07 18:01:29,010][riot01][DEBUG ] "mon": [ -[2020-03-07 18:01:29,010][riot01][DEBUG ] { -[2020-03-07 18:01:29,010][riot01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-07 18:01:29,010][riot01][DEBUG ] "num": 1, -[2020-03-07 18:01:29,011][riot01][DEBUG ] "release": "luminous" -[2020-03-07 18:01:29,011][riot01][DEBUG ] } -[2020-03-07 18:01:29,011][riot01][DEBUG ] ] -[2020-03-07 18:01:29,011][riot01][DEBUG ] }, -[2020-03-07 18:01:29,011][riot01][DEBUG ] "features": { -[2020-03-07 18:01:29,012][riot01][DEBUG ] "quorum_con": "0", -[2020-03-07 18:01:29,012][riot01][DEBUG ] "quorum_mon": [], -[2020-03-07 18:01:29,012][riot01][DEBUG ] "required_con": "0", -[2020-03-07 18:01:29,013][riot01][DEBUG ] "required_mon": [] -[2020-03-07 18:01:29,013][riot01][DEBUG ] }, -[2020-03-07 18:01:29,013][riot01][DEBUG ] "monmap": { -[2020-03-07 18:01:29,013][riot01][DEBUG ] "created": "2020-03-07 18:01:19.539212", -[2020-03-07 18:01:29,013][riot01][DEBUG ] "epoch": 0, -[2020-03-07 18:01:29,013][riot01][DEBUG ] "features": { -[2020-03-07 18:01:29,014][riot01][DEBUG ] "optional": [], -[2020-03-07 18:01:29,014][riot01][DEBUG ] "persistent": [] -[2020-03-07 18:01:29,014][riot01][DEBUG ] }, -[2020-03-07 18:01:29,014][riot01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-07 18:01:29,015][riot01][DEBUG ] "min_mon_release": 0, -[2020-03-07 18:01:29,015][riot01][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-07 18:01:29,015][riot01][DEBUG ] "modified": "2020-03-07 18:01:19.539212", -[2020-03-07 18:01:29,015][riot01][DEBUG ] "mons": [ -[2020-03-07 18:01:29,016][riot01][DEBUG ] { -[2020-03-07 18:01:29,016][riot01][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-07 18:01:29,016][riot01][DEBUG ] "name": "pine01", -[2020-03-07 18:01:29,016][riot01][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-07 18:01:29,017][riot01][DEBUG ] "public_addrs": { -[2020-03-07 18:01:29,017][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:01:29,017][riot01][DEBUG ] { -[2020-03-07 18:01:29,017][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-07 18:01:29,017][riot01][DEBUG ] "nonce": 1, -[2020-03-07 18:01:29,017][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:01:29,018][riot01][DEBUG ] } -[2020-03-07 18:01:29,018][riot01][DEBUG ] ] -[2020-03-07 18:01:29,018][riot01][DEBUG ] }, -[2020-03-07 18:01:29,018][riot01][DEBUG ] "rank": 0 -[2020-03-07 18:01:29,018][riot01][DEBUG ] }, -[2020-03-07 18:01:29,019][riot01][DEBUG ] { -[2020-03-07 18:01:29,019][riot01][DEBUG ] "addr": "0.0.0.0:0/2", -[2020-03-07 18:01:29,019][riot01][DEBUG ] "name": "pine02", -[2020-03-07 18:01:29,019][riot01][DEBUG ] "public_addr": "0.0.0.0:0/2", -[2020-03-07 18:01:29,019][riot01][DEBUG ] "public_addrs": { -[2020-03-07 18:01:29,019][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:01:29,020][riot01][DEBUG ] { -[2020-03-07 18:01:29,020][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-07 18:01:29,020][riot01][DEBUG ] "nonce": 2, -[2020-03-07 18:01:29,020][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:01:29,020][riot01][DEBUG ] } -[2020-03-07 18:01:29,020][riot01][DEBUG ] ] -[2020-03-07 18:01:29,021][riot01][DEBUG ] }, -[2020-03-07 18:01:29,021][riot01][DEBUG ] "rank": 1 -[2020-03-07 18:01:29,021][riot01][DEBUG ] } -[2020-03-07 18:01:29,021][riot01][DEBUG ] ] -[2020-03-07 18:01:29,021][riot01][DEBUG ] }, -[2020-03-07 18:01:29,021][riot01][DEBUG ] "name": "riot01", -[2020-03-07 18:01:29,022][riot01][DEBUG ] "outside_quorum": [], -[2020-03-07 18:01:29,022][riot01][DEBUG ] "quorum": [], -[2020-03-07 18:01:29,022][riot01][DEBUG ] "rank": -1, -[2020-03-07 18:01:29,022][riot01][DEBUG ] "state": "probing", -[2020-03-07 18:01:29,022][riot01][DEBUG ] "sync_provider": [] -[2020-03-07 18:01:29,023][riot01][DEBUG ] } -[2020-03-07 18:01:29,023][riot01][DEBUG ] ******************************************************************************** -[2020-03-07 18:01:29,023][riot01][INFO ] monitor: mon.riot01 is currently at the state of probing -[2020-03-07 18:01:29,042][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-07 18:01:30,062][riot01][WARNING] riot01 is not defined in `mon initial members` -[2020-03-07 18:01:30,062][riot01][WARNING] monitor riot01 does not exist in monmap -[2020-03-07 18:08:11,337][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] username : root -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 18:08:11,337][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 18:08:11,338][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-07 18:08:11,338][ceph_deploy.cli][INFO ] func : -[2020-03-07 18:08:11,338][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 18:08:11,338][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 18:08:11,338][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-07 18:08:12,803][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 18:08:12,804][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 18:08:12,975][riot01][DEBUG ] detect machine type -[2020-03-07 18:08:13,009][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:08:13,029][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-07 18:08:14,460][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 18:08:14,460][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 18:08:14,579][pine01][DEBUG ] detect machine type -[2020-03-07 18:08:14,694][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:08:14,920][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-07 18:08:16,450][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-07 18:08:16,450][pine02][DEBUG ] detect platform information from remote host -[2020-03-07 18:08:16,573][pine02][DEBUG ] detect machine type -[2020-03-07 18:08:16,600][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:08:16,612][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-07 18:08:18,833][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-07 18:08:18,835][ebin01][DEBUG ] detect platform information from remote host -[2020-03-07 18:08:19,044][ebin01][DEBUG ] detect machine type -[2020-03-07 18:08:19,075][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:08:19,097][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-07 18:08:22,348][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-07 18:08:22,350][ebin02][DEBUG ] detect platform information from remote host -[2020-03-07 18:08:22,645][ebin02][DEBUG ] detect machine type -[2020-03-07 18:08:22,689][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:08:22,715][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-07 18:08:23,268][lenny][DEBUG ] connected to host: root@lenny -[2020-03-07 18:08:23,268][lenny][DEBUG ] detect platform information from remote host -[2020-03-07 18:08:23,292][lenny][DEBUG ] detect machine type -[2020-03-07 18:08:23,295][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:10:14,304][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf mon create riot01 -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] username : root -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] mon : ['riot01'] -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] func : -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] keyrings : None -[2020-03-07 18:10:14,304][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 18:10:14,305][ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts riot01 -[2020-03-07 18:10:14,305][ceph_deploy.mon][DEBUG ] detecting platform for host riot01 ... -[2020-03-07 18:10:15,798][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 18:10:15,800][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 18:10:15,969][riot01][DEBUG ] detect machine type -[2020-03-07 18:10:16,007][riot01][DEBUG ] find the location of an executable -[2020-03-07 18:10:16,014][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-07 18:10:16,015][riot01][DEBUG ] determining if provided host has same hostname in remote -[2020-03-07 18:10:16,015][riot01][DEBUG ] get remote short hostname -[2020-03-07 18:10:16,021][riot01][DEBUG ] deploying mon to riot01 -[2020-03-07 18:10:16,021][riot01][DEBUG ] get remote short hostname -[2020-03-07 18:10:16,027][riot01][DEBUG ] remote hostname: riot01 -[2020-03-07 18:10:16,040][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 18:10:16,062][riot01][DEBUG ] create the mon path if it does not exist -[2020-03-07 18:10:16,069][riot01][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-riot01/done -[2020-03-07 18:10:16,075][riot01][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-07 18:10:16,082][riot01][DEBUG ] create the init path if it does not exist -[2020-03-07 18:10:16,097][riot01][INFO ] Running command: systemctl enable ceph.target -[2020-03-07 18:10:18,337][riot01][INFO ] Running command: systemctl enable ceph-mon@riot01 -[2020-03-07 18:10:20,581][riot01][INFO ] Running command: systemctl start ceph-mon@riot01 -[2020-03-07 18:10:22,721][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-07 18:10:23,591][riot01][DEBUG ] ******************************************************************************** -[2020-03-07 18:10:23,591][riot01][DEBUG ] status for monitor: mon.riot01 -[2020-03-07 18:10:23,593][riot01][DEBUG ] { -[2020-03-07 18:10:23,593][riot01][DEBUG ] "election_epoch": 0, -[2020-03-07 18:10:23,593][riot01][DEBUG ] "extra_probe_peers": [ -[2020-03-07 18:10:23,593][riot01][DEBUG ] { -[2020-03-07 18:10:23,594][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:10:23,594][riot01][DEBUG ] { -[2020-03-07 18:10:23,594][riot01][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-07 18:10:23,594][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:10:23,594][riot01][DEBUG ] "type": "v2" -[2020-03-07 18:10:23,595][riot01][DEBUG ] }, -[2020-03-07 18:10:23,595][riot01][DEBUG ] { -[2020-03-07 18:10:23,595][riot01][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-07 18:10:23,595][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:10:23,595][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:10:23,595][riot01][DEBUG ] } -[2020-03-07 18:10:23,596][riot01][DEBUG ] ] -[2020-03-07 18:10:23,596][riot01][DEBUG ] }, -[2020-03-07 18:10:23,596][riot01][DEBUG ] { -[2020-03-07 18:10:23,596][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:10:23,596][riot01][DEBUG ] { -[2020-03-07 18:10:23,596][riot01][DEBUG ] "addr": "192.168.10.161:3300", -[2020-03-07 18:10:23,597][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:10:23,597][riot01][DEBUG ] "type": "v2" -[2020-03-07 18:10:23,597][riot01][DEBUG ] }, -[2020-03-07 18:10:23,597][riot01][DEBUG ] { -[2020-03-07 18:10:23,597][riot01][DEBUG ] "addr": "192.168.10.161:6789", -[2020-03-07 18:10:23,597][riot01][DEBUG ] "nonce": 0, -[2020-03-07 18:10:23,598][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:10:23,598][riot01][DEBUG ] } -[2020-03-07 18:10:23,598][riot01][DEBUG ] ] -[2020-03-07 18:10:23,598][riot01][DEBUG ] } -[2020-03-07 18:10:23,598][riot01][DEBUG ] ], -[2020-03-07 18:10:23,598][riot01][DEBUG ] "feature_map": { -[2020-03-07 18:10:23,599][riot01][DEBUG ] "mon": [ -[2020-03-07 18:10:23,599][riot01][DEBUG ] { -[2020-03-07 18:10:23,599][riot01][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-07 18:10:23,599][riot01][DEBUG ] "num": 1, -[2020-03-07 18:10:23,600][riot01][DEBUG ] "release": "luminous" -[2020-03-07 18:10:23,600][riot01][DEBUG ] } -[2020-03-07 18:10:23,600][riot01][DEBUG ] ] -[2020-03-07 18:10:23,600][riot01][DEBUG ] }, -[2020-03-07 18:10:23,600][riot01][DEBUG ] "features": { -[2020-03-07 18:10:23,601][riot01][DEBUG ] "quorum_con": "0", -[2020-03-07 18:10:23,601][riot01][DEBUG ] "quorum_mon": [], -[2020-03-07 18:10:23,601][riot01][DEBUG ] "required_con": "0", -[2020-03-07 18:10:23,601][riot01][DEBUG ] "required_mon": [] -[2020-03-07 18:10:23,601][riot01][DEBUG ] }, -[2020-03-07 18:10:23,601][riot01][DEBUG ] "monmap": { -[2020-03-07 18:10:23,602][riot01][DEBUG ] "created": "2020-03-07 18:01:19.539212", -[2020-03-07 18:10:23,602][riot01][DEBUG ] "epoch": 0, -[2020-03-07 18:10:23,602][riot01][DEBUG ] "features": { -[2020-03-07 18:10:23,602][riot01][DEBUG ] "optional": [], -[2020-03-07 18:10:23,602][riot01][DEBUG ] "persistent": [] -[2020-03-07 18:10:23,602][riot01][DEBUG ] }, -[2020-03-07 18:10:23,603][riot01][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-07 18:10:23,603][riot01][DEBUG ] "min_mon_release": 0, -[2020-03-07 18:10:23,603][riot01][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-07 18:10:23,603][riot01][DEBUG ] "modified": "2020-03-07 18:01:19.539212", -[2020-03-07 18:10:23,603][riot01][DEBUG ] "mons": [ -[2020-03-07 18:10:23,603][riot01][DEBUG ] { -[2020-03-07 18:10:23,604][riot01][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-07 18:10:23,604][riot01][DEBUG ] "name": "pine01", -[2020-03-07 18:10:23,604][riot01][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-07 18:10:23,604][riot01][DEBUG ] "public_addrs": { -[2020-03-07 18:10:23,604][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:10:23,605][riot01][DEBUG ] { -[2020-03-07 18:10:23,605][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-07 18:10:23,605][riot01][DEBUG ] "nonce": 1, -[2020-03-07 18:10:23,605][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:10:23,605][riot01][DEBUG ] } -[2020-03-07 18:10:23,605][riot01][DEBUG ] ] -[2020-03-07 18:10:23,606][riot01][DEBUG ] }, -[2020-03-07 18:10:23,606][riot01][DEBUG ] "rank": 0 -[2020-03-07 18:10:23,606][riot01][DEBUG ] }, -[2020-03-07 18:10:23,607][riot01][DEBUG ] { -[2020-03-07 18:10:23,607][riot01][DEBUG ] "addr": "0.0.0.0:0/2", -[2020-03-07 18:10:23,607][riot01][DEBUG ] "name": "pine02", -[2020-03-07 18:10:23,607][riot01][DEBUG ] "public_addr": "0.0.0.0:0/2", -[2020-03-07 18:10:23,607][riot01][DEBUG ] "public_addrs": { -[2020-03-07 18:10:23,608][riot01][DEBUG ] "addrvec": [ -[2020-03-07 18:10:23,608][riot01][DEBUG ] { -[2020-03-07 18:10:23,608][riot01][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-07 18:10:23,609][riot01][DEBUG ] "nonce": 2, -[2020-03-07 18:10:23,609][riot01][DEBUG ] "type": "v1" -[2020-03-07 18:10:23,609][riot01][DEBUG ] } -[2020-03-07 18:10:23,609][riot01][DEBUG ] ] -[2020-03-07 18:10:23,609][riot01][DEBUG ] }, -[2020-03-07 18:10:23,610][riot01][DEBUG ] "rank": 1 -[2020-03-07 18:10:23,610][riot01][DEBUG ] } -[2020-03-07 18:10:23,610][riot01][DEBUG ] ] -[2020-03-07 18:10:23,610][riot01][DEBUG ] }, -[2020-03-07 18:10:23,610][riot01][DEBUG ] "name": "riot01", -[2020-03-07 18:10:23,611][riot01][DEBUG ] "outside_quorum": [], -[2020-03-07 18:10:23,611][riot01][DEBUG ] "quorum": [], -[2020-03-07 18:10:23,611][riot01][DEBUG ] "rank": -1, -[2020-03-07 18:10:23,611][riot01][DEBUG ] "state": "probing", -[2020-03-07 18:10:23,611][riot01][DEBUG ] "sync_provider": [] -[2020-03-07 18:10:23,611][riot01][DEBUG ] } -[2020-03-07 18:10:23,612][riot01][DEBUG ] ******************************************************************************** -[2020-03-07 18:10:23,612][riot01][INFO ] monitor: mon.riot01 is currently at the state of probing -[2020-03-07 18:10:23,627][riot01][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.riot01.asok mon_status -[2020-03-07 18:10:24,498][riot01][WARNING] monitor riot01 does not exist in monmap -[2020-03-07 18:52:37,516][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin riot01 -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] username : root -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] client : ['riot01'] -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] func : -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 18:52:37,517][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 18:52:37,517][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-07 18:52:39,006][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 18:52:39,009][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 18:52:39,178][riot01][DEBUG ] detect machine type -[2020-03-07 18:52:39,221][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:06:04,940][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 19:06:04,940][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-07 19:06:04,940][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 19:06:04,940][ceph_deploy.cli][INFO ] username : root -[2020-03-07 19:06:04,940][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] func : -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 19:06:04,941][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 19:06:04,941][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-07 19:06:06,415][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 19:06:06,418][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 19:06:06,587][riot01][DEBUG ] detect machine type -[2020-03-07 19:06:06,625][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:06:06,647][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-07 19:06:13,850][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 19:06:13,852][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 19:06:14,022][pine01][DEBUG ] detect machine type -[2020-03-07 19:06:14,222][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:06:14,232][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-07 19:06:15,840][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-07 19:06:15,842][pine02][DEBUG ] detect platform information from remote host -[2020-03-07 19:06:15,962][pine02][DEBUG ] detect machine type -[2020-03-07 19:06:15,987][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:06:15,997][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-07 19:06:18,553][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-07 19:06:18,555][ebin01][DEBUG ] detect platform information from remote host -[2020-03-07 19:06:18,768][ebin01][DEBUG ] detect machine type -[2020-03-07 19:06:18,802][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:06:18,824][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-07 19:06:22,141][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-07 19:06:22,143][ebin02][DEBUG ] detect platform information from remote host -[2020-03-07 19:06:22,408][ebin02][DEBUG ] detect machine type -[2020-03-07 19:06:22,450][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:06:22,474][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-07 19:06:22,892][lenny][DEBUG ] connected to host: root@lenny -[2020-03-07 19:06:22,893][lenny][DEBUG ] detect platform information from remote host -[2020-03-07 19:06:22,909][lenny][DEBUG ] detect machine type -[2020-03-07 19:06:22,912][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:06:59,571][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 19:06:59,571][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-07 19:06:59,571][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 19:06:59,571][ceph_deploy.cli][INFO ] username : root -[2020-03-07 19:06:59,571][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 19:06:59,571][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 19:06:59,571][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-07 19:06:59,571][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 19:06:59,572][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 19:06:59,572][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 19:06:59,572][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-07 19:06:59,572][ceph_deploy.cli][INFO ] func : -[2020-03-07 19:06:59,572][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 19:06:59,572][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 19:06:59,572][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-07 19:07:01,089][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 19:07:01,091][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 19:07:01,260][riot01][DEBUG ] detect machine type -[2020-03-07 19:07:01,304][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:07:01,326][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-07 19:07:08,801][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 19:07:08,803][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 19:07:09,080][pine01][DEBUG ] detect machine type -[2020-03-07 19:07:09,381][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:07:09,524][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-07 19:07:11,083][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-07 19:07:11,085][pine02][DEBUG ] detect platform information from remote host -[2020-03-07 19:07:11,207][pine02][DEBUG ] detect machine type -[2020-03-07 19:07:11,235][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:07:11,247][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-07 19:07:13,489][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-07 19:07:13,491][ebin01][DEBUG ] detect platform information from remote host -[2020-03-07 19:07:13,654][ebin01][DEBUG ] detect machine type -[2020-03-07 19:07:13,685][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:07:13,699][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-07 19:07:16,385][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-07 19:07:16,387][ebin02][DEBUG ] detect platform information from remote host -[2020-03-07 19:07:16,548][ebin02][DEBUG ] detect machine type -[2020-03-07 19:07:16,580][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:07:16,594][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-07 19:07:16,964][lenny][DEBUG ] connected to host: root@lenny -[2020-03-07 19:07:16,964][lenny][DEBUG ] detect platform information from remote host -[2020-03-07 19:07:16,980][lenny][DEBUG ] detect machine type -[2020-03-07 19:07:16,984][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:58:48,850][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 19:58:48,850][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny -[2020-03-07 19:58:48,850][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 19:58:48,850][ceph_deploy.cli][INFO ] username : root -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny'] -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] func : -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 19:58:48,851][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 19:58:48,851][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-07 19:58:50,337][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 19:58:50,340][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 19:58:50,509][riot01][DEBUG ] detect machine type -[2020-03-07 19:58:50,556][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:58:50,580][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-07 19:58:56,714][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 19:58:56,716][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 19:58:56,879][pine01][DEBUG ] detect machine type -[2020-03-07 19:58:57,068][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:58:57,126][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-07 19:58:58,729][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-07 19:58:58,731][pine02][DEBUG ] detect platform information from remote host -[2020-03-07 19:58:58,852][pine02][DEBUG ] detect machine type -[2020-03-07 19:58:58,880][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:58:58,893][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-07 19:59:01,207][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-07 19:59:01,209][ebin01][DEBUG ] detect platform information from remote host -[2020-03-07 19:59:01,370][ebin01][DEBUG ] detect machine type -[2020-03-07 19:59:01,403][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:59:01,419][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-07 19:59:04,148][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-07 19:59:04,150][ebin02][DEBUG ] detect platform information from remote host -[2020-03-07 19:59:04,313][ebin02][DEBUG ] detect machine type -[2020-03-07 19:59:04,345][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 19:59:04,360][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-07 19:59:04,799][lenny][DEBUG ] connected to host: root@lenny -[2020-03-07 19:59:04,800][lenny][DEBUG ] detect platform information from remote host -[2020-03-07 19:59:04,815][lenny][DEBUG ] detect machine type -[2020-03-07 19:59:04,819][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 20:00:01,161][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create riot01 -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] username : root -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] rgw : [('riot01', 'rgw.riot01')] -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] func : -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 20:00:01,161][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 20:00:01,162][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts riot01:rgw.riot01 -[2020-03-07 20:00:02,623][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 20:00:02,625][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 20:00:02,793][riot01][DEBUG ] detect machine type -[2020-03-07 20:00:02,829][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-07 20:00:02,829][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-07 20:00:02,830][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to riot01 -[2020-03-07 20:00:02,830][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 20:00:02,856][riot01][WARNING] rgw keyring does not exist yet, creating one -[2020-03-07 20:00:02,857][riot01][DEBUG ] create a keyring file -[2020-03-07 20:00:02,871][riot01][DEBUG ] create path recursively if it doesn't exist -[2020-03-07 20:00:02,889][riot01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.riot01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.riot01/keyring -[2020-03-07 20:00:03,770][riot01][ERROR ] Traceback (most recent call last): -[2020-03-07 20:00:03,771][riot01][ERROR ] File "/usr/bin/ceph", line 1266, in -[2020-03-07 20:00:03,771][riot01][ERROR ] retval = main() -[2020-03-07 20:00:03,772][riot01][ERROR ] File "/usr/bin/ceph", line 979, in main -[2020-03-07 20:00:03,772][riot01][ERROR ] conffile=conffile) -[2020-03-07 20:00:03,772][riot01][ERROR ] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-07 20:00:03,772][riot01][ERROR ] raise Exception("timed out") -[2020-03-07 20:00:03,773][riot01][ERROR ] Exception: timed out -[2020-03-07 20:00:03,773][riot01][ERROR ] exit code from command was: 1 -[2020-03-07 20:00:03,773][ceph_deploy.rgw][ERROR ] could not create rgw -[2020-03-07 20:00:03,773][ceph_deploy][ERROR ] GenericError: Failed to create 1 RGWs - -[2020-03-07 20:00:12,282][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin riot01 -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] username : root -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] client : ['riot01'] -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] func : -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 20:00:12,283][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 20:00:12,283][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to riot01 -[2020-03-07 20:00:13,814][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 20:00:13,816][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 20:00:13,985][riot01][DEBUG ] detect machine type -[2020-03-07 20:00:14,022][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 20:00:16,579][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 20:00:16,579][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create riot01 -[2020-03-07 20:00:16,579][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 20:00:16,579][ceph_deploy.cli][INFO ] username : root -[2020-03-07 20:00:16,579][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 20:00:16,579][ceph_deploy.cli][INFO ] rgw : [('riot01', 'rgw.riot01')] -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] func : -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 20:00:16,580][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 20:00:16,580][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts riot01:rgw.riot01 -[2020-03-07 20:00:18,117][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-07 20:00:18,119][riot01][DEBUG ] detect platform information from remote host -[2020-03-07 20:00:18,288][riot01][DEBUG ] detect machine type -[2020-03-07 20:00:18,328][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-07 20:00:18,328][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-07 20:00:18,330][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to riot01 -[2020-03-07 20:00:18,330][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 20:00:18,359][riot01][DEBUG ] create path recursively if it doesn't exist -[2020-03-07 20:00:18,376][riot01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.riot01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.riot01/keyring -[2020-03-07 20:00:19,257][riot01][ERROR ] Traceback (most recent call last): -[2020-03-07 20:00:19,261][riot01][ERROR ] File "/usr/bin/ceph", line 1266, in -[2020-03-07 20:00:19,262][riot01][ERROR ] retval = main() -[2020-03-07 20:00:19,262][riot01][ERROR ] File "/usr/bin/ceph", line 979, in main -[2020-03-07 20:00:19,262][riot01][ERROR ] conffile=conffile) -[2020-03-07 20:00:19,263][riot01][ERROR ] File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1319, in run_in_thread -[2020-03-07 20:00:19,263][riot01][ERROR ] raise Exception("timed out") -[2020-03-07 20:00:19,264][riot01][ERROR ] Exception: timed out -[2020-03-07 20:00:19,264][riot01][ERROR ] exit code from command was: 1 -[2020-03-07 20:00:19,264][ceph_deploy.rgw][ERROR ] could not create rgw -[2020-03-07 20:00:19,264][ceph_deploy][ERROR ] GenericError: Failed to create 1 RGWs - -[2020-03-07 20:00:23,025][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create pine01 -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] username : root -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] rgw : [('pine01', 'rgw.pine01')] -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] func : -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 20:00:23,026][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 20:00:23,026][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts pine01:rgw.pine01 -[2020-03-07 20:00:29,546][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-07 20:00:29,548][pine01][DEBUG ] detect platform information from remote host -[2020-03-07 20:00:29,750][pine01][DEBUG ] detect machine type -[2020-03-07 20:00:29,889][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-07 20:00:29,889][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-07 20:00:29,891][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to pine01 -[2020-03-07 20:00:29,891][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 20:00:29,961][pine01][WARNING] rgw keyring does not exist yet, creating one -[2020-03-07 20:00:29,961][pine01][DEBUG ] create a keyring file -[2020-03-07 20:00:30,393][pine01][DEBUG ] create path recursively if it doesn't exist -[2020-03-07 20:00:30,522][pine01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.pine01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.pine01/keyring -[2020-03-07 20:00:33,889][pine01][INFO ] Running command: systemctl enable ceph-radosgw@rgw.pine01 -[2020-03-07 20:00:35,063][pine01][INFO ] Running command: systemctl start ceph-radosgw@rgw.pine01 -[2020-03-07 20:00:35,286][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-07 20:00:36,214][ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host pine01 and default port 7480 -[2020-03-07 20:00:42,293][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create pine02 -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] username : root -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] rgw : [('pine02', 'rgw.pine02')] -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] func : -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 20:00:42,293][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 20:00:42,293][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts pine02:rgw.pine02 -[2020-03-07 20:00:43,873][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-07 20:00:43,876][pine02][DEBUG ] detect platform information from remote host -[2020-03-07 20:00:43,995][pine02][DEBUG ] detect machine type -[2020-03-07 20:00:44,023][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-07 20:00:44,024][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-07 20:00:44,026][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to pine02 -[2020-03-07 20:00:44,026][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 20:00:44,046][pine02][WARNING] rgw keyring does not exist yet, creating one -[2020-03-07 20:00:44,047][pine02][DEBUG ] create a keyring file -[2020-03-07 20:00:44,061][pine02][DEBUG ] create path recursively if it doesn't exist -[2020-03-07 20:00:44,082][pine02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.pine02 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.pine02/keyring -[2020-03-07 20:00:51,113][pine02][INFO ] Running command: systemctl enable ceph-radosgw@rgw.pine02 -[2020-03-07 20:00:51,800][pine02][INFO ] Running command: systemctl start ceph-radosgw@rgw.pine02 -[2020-03-07 20:00:51,883][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-07 20:00:52,709][ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host pine02 and default port 7480 -[2020-03-07 22:26:38,398][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin tumor.chaos -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] username : root -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 22:26:38,399][ceph_deploy.cli][INFO ] client : ['tumor.chaos'] -[2020-03-07 22:26:38,400][ceph_deploy.cli][INFO ] func : -[2020-03-07 22:26:38,400][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 22:26:38,400][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 22:26:38,400][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to tumor.chaos -[2020-03-07 22:26:39,552][tumor.chaos][DEBUG ] connected to host: root@tumor.chaos -[2020-03-07 22:26:39,554][tumor.chaos][DEBUG ] detect platform information from remote host -[2020-03-07 22:26:39,841][tumor.chaos][DEBUG ] detect machine type -[2020-03-07 22:26:39,862][tumor.chaos][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-07 22:27:25,305][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purge tumor.chaos -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] username : root -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] host : ['tumor.chaos'] -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] func : -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 22:27:25,306][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 22:27:25,306][ceph_deploy.install][INFO ] note that some dependencies *will not* be removed because they can cause issues with qemu-kvm -[2020-03-07 22:27:25,306][ceph_deploy.install][INFO ] like: librbd1 and librados2 -[2020-03-07 22:27:25,306][ceph_deploy.install][DEBUG ] Purging on cluster ceph hosts tumor.chaos -[2020-03-07 22:27:25,307][ceph_deploy.install][DEBUG ] Detecting platform for host tumor.chaos ... -[2020-03-07 22:27:26,297][tumor.chaos][DEBUG ] connected to host: root@tumor.chaos -[2020-03-07 22:27:26,299][tumor.chaos][DEBUG ] detect platform information from remote host -[2020-03-07 22:27:26,536][tumor.chaos][DEBUG ] detect machine type -[2020-03-07 22:27:26,556][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-07 22:27:26,557][tumor.chaos][INFO ] Purging Ceph on tumor.chaos -[2020-03-07 22:27:26,564][tumor.chaos][INFO ] Running command: env DEBIAN_FRONTEND=noninteractive DEBIAN_PRIORITY=critical apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw -[2020-03-07 22:27:27,152][tumor.chaos][DEBUG ] Reading package lists... -[2020-03-07 22:27:28,521][tumor.chaos][DEBUG ] Building dependency tree... -[2020-03-07 22:27:28,537][tumor.chaos][DEBUG ] Reading state information... -[2020-03-07 22:27:30,257][tumor.chaos][DEBUG ] Package 'ceph-fs-common' is not installed, so not removed -[2020-03-07 22:27:30,258][tumor.chaos][DEBUG ] Package 'ceph-mds' is not installed, so not removed -[2020-03-07 22:27:30,258][tumor.chaos][DEBUG ] Package 'radosgw' is not installed, so not removed -[2020-03-07 22:27:30,258][tumor.chaos][DEBUG ] The following packages were automatically installed and are no longer required: -[2020-03-07 22:27:30,259][tumor.chaos][DEBUG ] gdisk libboost-context1.67.0 libboost-coroutine1.67.0 libboost-python1.67.0 -[2020-03-07 22:27:30,259][tumor.chaos][DEBUG ] libboost-random1.67.0 libboost-regex1.67.0 libgoogle-perftools4 liboath0 -[2020-03-07 22:27:30,259][tumor.chaos][DEBUG ] librabbitmq4 libradosstriper1 libtcmalloc-minimal4 python3-bcrypt -[2020-03-07 22:27:30,259][tumor.chaos][DEBUG ] python3-bs4 python3-ceph-argparse python3-cephfs python3-cherrypy3 -[2020-03-07 22:27:30,260][tumor.chaos][DEBUG ] python3-jwt python3-logutils python3-paste python3-pastedeploy python3-pecan -[2020-03-07 22:27:30,260][tumor.chaos][DEBUG ] python3-prettytable python3-rados python3-rbd python3-simplegeneric -[2020-03-07 22:27:30,260][tumor.chaos][DEBUG ] python3-singledispatch python3-soupsieve python3-tempita python3-waitress -[2020-03-07 22:27:30,260][tumor.chaos][DEBUG ] python3-webob python3-webtest python3-werkzeug -[2020-03-07 22:27:30,260][tumor.chaos][DEBUG ] Use 'apt autoremove' to remove them. -[2020-03-07 22:27:30,676][tumor.chaos][DEBUG ] The following packages will be REMOVED: -[2020-03-07 22:27:30,677][tumor.chaos][DEBUG ] ceph* ceph-base* ceph-common* ceph-mgr* ceph-mon* ceph-osd* -[2020-03-07 22:27:33,150][tumor.chaos][DEBUG ] [master 495f25a] saving uncommitted changes in /etc prior to apt run -[2020-03-07 22:27:33,151][tumor.chaos][DEBUG ] 4 files changed, 36 insertions(+) -[2020-03-07 22:27:33,151][tumor.chaos][DEBUG ] create mode 100644 ceph/ceph.client.admin.keyring -[2020-03-07 22:27:33,151][tumor.chaos][DEBUG ] create mode 100644 ceph/ceph.conf -[2020-03-07 22:27:33,151][tumor.chaos][DEBUG ] create mode 100644 ceph/tmpo8ylPR -[2020-03-07 22:27:34,972][tumor.chaos][DEBUG ] 0 upgraded, 0 newly installed, 6 to remove and 3 not upgraded. -[2020-03-07 22:27:34,973][tumor.chaos][DEBUG ] After this operation, 144 MB disk space will be freed. -[2020-03-07 22:27:35,238][tumor.chaos][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 146703 files and directories currently installed.) -[2020-03-07 22:27:35,254][tumor.chaos][DEBUG ] Removing ceph (14.2.7-1~bpo10+1) ... -[2020-03-07 22:27:35,419][tumor.chaos][DEBUG ] Removing ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-07 22:27:39,050][tumor.chaos][DEBUG ] Removing ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-07 22:27:43,376][tumor.chaos][DEBUG ] Removing ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-07 22:27:47,051][tumor.chaos][DEBUG ] Removing ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-07 22:27:50,626][tumor.chaos][DEBUG ] Removing ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-07 22:27:54,505][tumor.chaos][DEBUG ] Processing triggers for libc-bin (2.28-10) ... -[2020-03-07 22:27:55,774][tumor.chaos][DEBUG ] (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 146298 files and directories currently installed.) -[2020-03-07 22:27:55,774][tumor.chaos][DEBUG ] Purging configuration files for ceph-mon (14.2.7-1~bpo10+1) ... -[2020-03-07 22:28:01,460][tumor.chaos][DEBUG ] Purging configuration files for ceph-base (14.2.7-1~bpo10+1) ... -[2020-03-07 22:28:06,844][tumor.chaos][DEBUG ] Purging configuration files for ceph (14.2.7-1~bpo10+1) ... -[2020-03-07 22:28:06,959][tumor.chaos][DEBUG ] Purging configuration files for ceph-mgr (14.2.7-1~bpo10+1) ... -[2020-03-07 22:28:12,292][tumor.chaos][DEBUG ] Purging configuration files for ceph-common (14.2.7-1~bpo10+1) ... -[2020-03-07 22:28:19,832][tumor.chaos][DEBUG ] Purging configuration files for ceph-osd (14.2.7-1~bpo10+1) ... -[2020-03-07 22:28:25,314][tumor.chaos][DEBUG ] Processing triggers for systemd (241-7~deb10u3) ... -[2020-03-07 22:28:29,998][tumor.chaos][DEBUG ] [master 10801ef] committing changes in /etc made by "apt-get --assume-yes -q -f --force-yes remove --purge ceph ceph-mds ceph-common ceph-fs-common radosgw" -[2020-03-07 22:28:29,999][tumor.chaos][DEBUG ] 27 files changed, 1 insertion(+), 144 deletions(-) -[2020-03-07 22:28:29,999][tumor.chaos][DEBUG ] delete mode 100644 ceph/ceph.client.admin.keyring -[2020-03-07 22:28:29,999][tumor.chaos][DEBUG ] delete mode 100644 ceph/ceph.conf -[2020-03-07 22:28:29,999][tumor.chaos][DEBUG ] delete mode 100644 ceph/rbdmap -[2020-03-07 22:28:30,000][tumor.chaos][DEBUG ] delete mode 100644 ceph/tmpo8ylPR -[2020-03-07 22:28:30,000][tumor.chaos][DEBUG ] delete mode 100644 default/ceph -[2020-03-07 22:28:30,000][tumor.chaos][DEBUG ] delete mode 100755 init.d/rbdmap -[2020-03-07 22:28:30,000][tumor.chaos][DEBUG ] delete mode 100644 logrotate.d/ceph-common -[2020-03-07 22:28:30,001][tumor.chaos][DEBUG ] delete mode 120000 rc0.d/K01rbdmap -[2020-03-07 22:28:30,001][tumor.chaos][DEBUG ] delete mode 120000 rc1.d/K01rbdmap -[2020-03-07 22:28:30,001][tumor.chaos][DEBUG ] delete mode 120000 rc2.d/S01rbdmap -[2020-03-07 22:28:30,002][tumor.chaos][DEBUG ] delete mode 120000 rc3.d/S01rbdmap -[2020-03-07 22:28:30,002][tumor.chaos][DEBUG ] delete mode 120000 rc4.d/S01rbdmap -[2020-03-07 22:28:30,002][tumor.chaos][DEBUG ] delete mode 120000 rc5.d/S01rbdmap -[2020-03-07 22:28:30,003][tumor.chaos][DEBUG ] delete mode 120000 rc6.d/K01rbdmap -[2020-03-07 22:28:30,003][tumor.chaos][DEBUG ] delete mode 100644 sudoers.d/ceph-osd-smartctl -[2020-03-07 22:28:30,003][tumor.chaos][DEBUG ] delete mode 100644 sysctl.d/30-ceph-osd.conf -[2020-03-07 22:28:30,003][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-crash.service -[2020-03-07 22:28:30,004][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mgr.target -[2020-03-07 22:28:30,004][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-mon.target -[2020-03-07 22:28:30,005][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/ceph.target.wants/ceph-osd.target -[2020-03-07 22:28:30,005][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mgr.target -[2020-03-07 22:28:30,006][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-mon.target -[2020-03-07 22:28:30,007][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph-osd.target -[2020-03-07 22:28:30,007][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/ceph.target -[2020-03-07 22:28:30,008][tumor.chaos][DEBUG ] delete mode 120000 systemd/system/multi-user.target.wants/rbdmap.service -[2020-03-07 22:28:30,175][tumor.chaos][WARNING] W: --force-yes is deprecated, use one of the options starting with --allow instead. -[2020-03-07 22:39:00,315][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root purgedata tumor -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] username : root -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] host : ['tumor'] -[2020-03-07 22:39:00,316][ceph_deploy.cli][INFO ] func : -[2020-03-07 22:39:00,317][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 22:39:00,317][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 22:39:00,317][ceph_deploy.install][DEBUG ] Purging data from cluster ceph hosts tumor -[2020-03-07 22:39:01,349][tumor][DEBUG ] connected to host: root@tumor -[2020-03-07 22:39:01,351][tumor][DEBUG ] detect platform information from remote host -[2020-03-07 22:39:01,532][tumor][DEBUG ] detect machine type -[2020-03-07 22:39:01,562][tumor][DEBUG ] find the location of an executable -[2020-03-07 22:39:02,542][tumor][DEBUG ] connected to host: root@tumor -[2020-03-07 22:39:02,543][tumor][DEBUG ] detect platform information from remote host -[2020-03-07 22:39:02,863][tumor][DEBUG ] detect machine type -[2020-03-07 22:39:02,886][ceph_deploy.install][INFO ] Distro info: debian 10.3 buster -[2020-03-07 22:39:02,886][tumor][INFO ] purging data on tumor -[2020-03-07 22:39:02,893][tumor][INFO ] Running command: rm -rf --one-file-system -- /var/lib/ceph -[2020-03-07 22:39:02,953][tumor][INFO ] Running command: rm -rf --one-file-system -- /etc/ceph/ -[2020-03-07 22:40:20,773][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-07 22:40:20,774][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin tumor.chaos -[2020-03-07 22:40:20,774][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] username : root -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] verbose : False -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] quiet : False -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] client : ['tumor.chaos'] -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] func : -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-07 22:40:20,775][ceph_deploy.cli][INFO ] default_release : False -[2020-03-07 22:40:20,775][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to tumor.chaos -[2020-03-07 22:40:21,757][tumor.chaos][DEBUG ] connected to host: root@tumor.chaos -[2020-03-07 22:40:21,759][tumor.chaos][DEBUG ] detect platform information from remote host -[2020-03-07 22:40:21,917][tumor.chaos][DEBUG ] detect machine type -[2020-03-07 22:40:21,938][tumor.chaos][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-08 19:38:41,142][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-08 19:38:41,152][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mds create pine01 pine02 -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] username : root -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] verbose : False -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] quiet : False -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] func : -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] mds : [('pine01', 'pine01'), ('pine02', 'pine02')] -[2020-03-08 19:38:41,153][ceph_deploy.cli][INFO ] default_release : False -[2020-03-08 19:38:41,154][ceph_deploy.mds][DEBUG ] Deploying mds, cluster ceph hosts pine01:pine01 pine02:pine02 -[2020-03-08 19:38:49,717][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-08 19:38:49,767][pine01][DEBUG ] detect platform information from remote host -[2020-03-08 19:38:50,995][pine01][DEBUG ] detect machine type -[2020-03-08 19:38:51,018][ceph_deploy.mds][INFO ] Distro info: debian 10.3 buster -[2020-03-08 19:38:51,018][ceph_deploy.mds][DEBUG ] remote host will use systemd -[2020-03-08 19:38:51,020][ceph_deploy.mds][DEBUG ] deploying mds bootstrap to pine01 -[2020-03-08 19:38:51,020][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-08 19:38:51,403][pine01][WARNING] mds keyring does not exist yet, creating one -[2020-03-08 19:38:51,403][pine01][DEBUG ] create a keyring file -[2020-03-08 19:38:51,417][pine01][DEBUG ] create path if it doesn't exist -[2020-03-08 19:38:51,435][pine01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mds --keyring /var/lib/ceph/bootstrap-mds/ceph.keyring auth get-or-create mds.pine01 osd allow rwx mds allow mon allow profile mds -o /var/lib/ceph/mds/ceph-pine01/keyring -[2020-03-08 19:38:54,878][pine01][INFO ] Running command: systemctl enable ceph-mds@pine01 -[2020-03-08 19:38:55,051][pine01][WARNING] Created symlink /etc/systemd/system/ceph-mds.target.wants/ceph-mds@pine01.service → /lib/systemd/system/ceph-mds@.service. -[2020-03-08 19:38:55,892][pine01][INFO ] Running command: systemctl start ceph-mds@pine01 -[2020-03-08 19:38:56,173][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-08 19:38:58,733][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-08 19:38:58,735][pine02][DEBUG ] detect platform information from remote host -[2020-03-08 19:38:58,892][pine02][DEBUG ] detect machine type -[2020-03-08 19:38:58,920][ceph_deploy.mds][INFO ] Distro info: debian 10.3 buster -[2020-03-08 19:38:58,921][ceph_deploy.mds][DEBUG ] remote host will use systemd -[2020-03-08 19:38:58,921][ceph_deploy.mds][DEBUG ] deploying mds bootstrap to pine02 -[2020-03-08 19:38:58,921][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-08 19:38:58,941][pine02][WARNING] mds keyring does not exist yet, creating one -[2020-03-08 19:38:58,941][pine02][DEBUG ] create a keyring file -[2020-03-08 19:38:58,955][pine02][DEBUG ] create path if it doesn't exist -[2020-03-08 19:38:58,974][pine02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mds --keyring /var/lib/ceph/bootstrap-mds/ceph.keyring auth get-or-create mds.pine02 osd allow rwx mds allow mon allow profile mds -o /var/lib/ceph/mds/ceph-pine02/keyring -[2020-03-08 19:39:02,728][pine02][INFO ] Running command: systemctl enable ceph-mds@pine02 -[2020-03-08 19:39:02,801][pine02][WARNING] Created symlink /etc/systemd/system/ceph-mds.target.wants/ceph-mds@pine02.service → /lib/systemd/system/ceph-mds@.service. -[2020-03-08 19:39:03,477][pine02][INFO ] Running command: systemctl start ceph-mds@pine02 -[2020-03-08 19:39:03,560][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-08 20:10:31,430][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mds create pine01 -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] username : root -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] verbose : False -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] quiet : False -[2020-03-08 20:10:31,430][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-08 20:10:31,431][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-08 20:10:31,431][ceph_deploy.cli][INFO ] func : -[2020-03-08 20:10:31,431][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-08 20:10:31,431][ceph_deploy.cli][INFO ] mds : [('pine01', 'pine01')] -[2020-03-08 20:10:31,431][ceph_deploy.cli][INFO ] default_release : False -[2020-03-08 20:10:31,431][ceph_deploy.mds][DEBUG ] Deploying mds, cluster ceph hosts pine01:pine01 -[2020-03-08 20:10:38,053][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-08 20:10:38,053][pine01][DEBUG ] detect platform information from remote host -[2020-03-08 20:10:38,225][pine01][DEBUG ] detect machine type -[2020-03-08 20:10:38,339][ceph_deploy.mds][INFO ] Distro info: debian 10.3 buster -[2020-03-08 20:10:38,339][ceph_deploy.mds][DEBUG ] remote host will use systemd -[2020-03-08 20:10:38,339][ceph_deploy.mds][DEBUG ] deploying mds bootstrap to pine01 -[2020-03-08 20:10:38,339][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-08 20:10:38,417][pine01][DEBUG ] create path if it doesn't exist -[2020-03-08 20:10:38,479][pine01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mds --keyring /var/lib/ceph/bootstrap-mds/ceph.keyring auth get-or-create mds.pine01 osd allow rwx mds allow mon allow profile mds -o /var/lib/ceph/mds/ceph-pine01/keyring -[2020-03-08 20:10:41,818][pine01][INFO ] Running command: systemctl enable ceph-mds@pine01 -[2020-03-08 20:10:42,742][pine01][INFO ] Running command: systemctl start ceph-mds@pine01 -[2020-03-08 20:10:43,075][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-08 20:14:14,144][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-08 20:14:14,144][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mds create pine02 -[2020-03-08 20:14:14,144][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] username : root -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] verbose : False -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] quiet : False -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] func : -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] mds : [('pine02', 'pine02')] -[2020-03-08 20:14:14,145][ceph_deploy.cli][INFO ] default_release : False -[2020-03-08 20:14:14,145][ceph_deploy.mds][DEBUG ] Deploying mds, cluster ceph hosts pine02:pine02 -[2020-03-08 20:14:15,675][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-08 20:14:15,676][pine02][DEBUG ] detect platform information from remote host -[2020-03-08 20:14:15,804][pine02][DEBUG ] detect machine type -[2020-03-08 20:14:15,828][ceph_deploy.mds][INFO ] Distro info: debian 10.3 buster -[2020-03-08 20:14:15,829][ceph_deploy.mds][DEBUG ] remote host will use systemd -[2020-03-08 20:14:15,829][ceph_deploy.mds][DEBUG ] deploying mds bootstrap to pine02 -[2020-03-08 20:14:15,829][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-08 20:14:15,848][pine02][DEBUG ] create path if it doesn't exist -[2020-03-08 20:14:15,861][pine02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mds --keyring /var/lib/ceph/bootstrap-mds/ceph.keyring auth get-or-create mds.pine02 osd allow rwx mds allow mon allow profile mds -o /var/lib/ceph/mds/ceph-pine02/keyring -[2020-03-08 20:14:18,907][pine02][INFO ] Running command: systemctl enable ceph-mds@pine02 -[2020-03-08 20:14:19,634][pine02][INFO ] Running command: systemctl start ceph-mds@pine02 -[2020-03-08 20:14:19,708][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-10 21:08:17,204][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create pine01 -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] username : root -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] verbose : False -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] rgw : [('pine01', 'rgw.pine01')] -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] quiet : False -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-10 21:08:17,215][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-10 21:08:17,216][ceph_deploy.cli][INFO ] func : -[2020-03-10 21:08:17,216][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-10 21:08:17,216][ceph_deploy.cli][INFO ] default_release : False -[2020-03-10 21:08:17,216][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts pine01:rgw.pine01 -[2020-03-10 21:08:25,594][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-10 21:08:25,597][pine01][DEBUG ] detect platform information from remote host -[2020-03-10 21:08:26,516][pine01][DEBUG ] detect machine type -[2020-03-10 21:08:26,628][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-10 21:08:26,628][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-10 21:08:26,630][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to pine01 -[2020-03-10 21:08:26,630][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-10 21:08:27,130][pine01][DEBUG ] create path recursively if it doesn't exist -[2020-03-10 21:08:27,209][pine01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.pine01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.pine01/keyring -[2020-03-10 21:08:30,770][pine01][INFO ] Running command: systemctl enable ceph-radosgw@rgw.pine01 -[2020-03-10 21:08:31,878][pine01][INFO ] Running command: systemctl start ceph-radosgw@rgw.pine01 -[2020-03-10 21:08:32,271][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-10 21:08:33,099][ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host pine01 and default port 7480 -[2020-03-10 22:05:08,473][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create pine02 -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] username : root -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] verbose : False -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] rgw : [('pine02', 'rgw.pine02')] -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] quiet : False -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] func : -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-10 22:05:08,473][ceph_deploy.cli][INFO ] default_release : False -[2020-03-10 22:05:08,473][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts pine02:rgw.pine02 -[2020-03-10 22:05:10,192][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-10 22:05:10,194][pine02][DEBUG ] detect platform information from remote host -[2020-03-10 22:05:10,363][pine02][DEBUG ] detect machine type -[2020-03-10 22:05:10,384][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-10 22:05:10,384][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-10 22:05:10,386][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to pine02 -[2020-03-10 22:05:10,386][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-10 22:05:10,407][pine02][DEBUG ] create path recursively if it doesn't exist -[2020-03-10 22:05:10,419][pine02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.pine02 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.pine02/keyring -[2020-03-10 22:05:13,675][pine02][INFO ] Running command: systemctl enable ceph-radosgw@rgw.pine02 -[2020-03-10 22:05:14,408][pine02][INFO ] Running command: systemctl start ceph-radosgw@rgw.pine02 -[2020-03-10 22:05:14,486][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-10 22:05:15,212][ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host pine02 and default port 7480 -[2020-03-10 22:13:21,593][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-10 22:13:21,593][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root rgw create pine01 -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] username : root -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] verbose : False -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] rgw : [('pine01', 'rgw.pine01')] -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] quiet : False -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] func : -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-10 22:13:21,594][ceph_deploy.cli][INFO ] default_release : False -[2020-03-10 22:13:21,594][ceph_deploy.rgw][DEBUG ] Deploying rgw, cluster ceph hosts pine01:rgw.pine01 -[2020-03-10 22:13:30,457][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-10 22:13:30,459][pine01][DEBUG ] detect platform information from remote host -[2020-03-10 22:13:31,219][pine01][DEBUG ] detect machine type -[2020-03-10 22:13:31,342][ceph_deploy.rgw][INFO ] Distro info: debian 10.3 buster -[2020-03-10 22:13:31,343][ceph_deploy.rgw][DEBUG ] remote host will use systemd -[2020-03-10 22:13:31,392][ceph_deploy.rgw][DEBUG ] deploying rgw bootstrap to pine01 -[2020-03-10 22:13:31,392][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-10 22:13:31,889][pine01][DEBUG ] create path recursively if it doesn't exist -[2020-03-10 22:13:31,958][pine01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-rgw --keyring /var/lib/ceph/bootstrap-rgw/ceph.keyring auth get-or-create client.rgw.pine01 osd allow rwx mon allow rw -o /var/lib/ceph/radosgw/ceph-rgw.pine01/keyring -[2020-03-10 22:13:35,484][pine01][INFO ] Running command: systemctl enable ceph-radosgw@rgw.pine01 -[2020-03-10 22:13:36,648][pine01][INFO ] Running command: systemctl start ceph-radosgw@rgw.pine01 -[2020-03-10 22:13:37,113][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-10 22:13:37,838][ceph_deploy.rgw][INFO ] The Ceph Object Gateway (RGW) is now running on host pine01 and default port 7480 -[2020-03-10 22:24:20,561][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-10 22:24:20,561][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin pine03 -[2020-03-10 22:24:20,561][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] username : root -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] verbose : False -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] quiet : False -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] client : ['pine03'] -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] func : -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-10 22:24:20,562][ceph_deploy.cli][INFO ] default_release : False -[2020-03-10 22:24:20,609][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to pine03 -[2020-03-10 22:24:31,355][pine03][DEBUG ] connected to host: root@pine03 -[2020-03-10 22:24:31,357][pine03][DEBUG ] detect platform information from remote host -[2020-03-10 22:24:32,583][pine03][DEBUG ] detect machine type -[2020-03-10 22:24:32,739][pine03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-10 22:27:11,130][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-10 22:27:11,130][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root admin tumor -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] username : root -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] verbose : False -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] quiet : False -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] client : ['tumor'] -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] func : -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-10 22:27:11,140][ceph_deploy.cli][INFO ] default_release : False -[2020-03-10 22:27:11,141][ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to tumor -[2020-03-10 22:27:12,447][tumor][DEBUG ] connected to host: root@tumor -[2020-03-10 22:27:12,449][tumor][DEBUG ] detect platform information from remote host -[2020-03-10 22:27:12,738][tumor][DEBUG ] detect machine type -[2020-03-10 22:27:12,758][tumor][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-11 20:51:02,058][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-11 20:51:02,060][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mds create pine02 -[2020-03-11 20:51:02,060][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-11 20:51:02,060][ceph_deploy.cli][INFO ] username : root -[2020-03-11 20:51:02,060][ceph_deploy.cli][INFO ] verbose : False -[2020-03-11 20:51:02,060][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-11 20:51:02,061][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-11 20:51:02,061][ceph_deploy.cli][INFO ] quiet : False -[2020-03-11 20:51:02,061][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-11 20:51:02,061][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-11 20:51:02,061][ceph_deploy.cli][INFO ] func : -[2020-03-11 20:51:02,061][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-11 20:51:02,062][ceph_deploy.cli][INFO ] mds : [('pine02', 'pine02')] -[2020-03-11 20:51:02,062][ceph_deploy.cli][INFO ] default_release : False -[2020-03-11 20:51:02,062][ceph_deploy.mds][DEBUG ] Deploying mds, cluster ceph hosts pine02:pine02 -[2020-03-11 20:51:03,998][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-11 20:51:04,010][pine02][DEBUG ] detect platform information from remote host -[2020-03-11 20:51:04,238][pine02][DEBUG ] detect machine type -[2020-03-11 20:51:04,256][ceph_deploy.mds][INFO ] Distro info: debian 10.3 buster -[2020-03-11 20:51:04,257][ceph_deploy.mds][DEBUG ] remote host will use systemd -[2020-03-11 20:51:04,258][ceph_deploy.mds][DEBUG ] deploying mds bootstrap to pine02 -[2020-03-11 20:51:04,259][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-11 20:51:04,276][pine02][DEBUG ] create path if it doesn't exist -[2020-03-11 20:51:04,289][pine02][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mds --keyring /var/lib/ceph/bootstrap-mds/ceph.keyring auth get-or-create mds.pine02 osd allow rwx mds allow mon allow profile mds -o /var/lib/ceph/mds/ceph-pine02/keyring -[2020-03-11 20:51:07,296][pine02][INFO ] Running command: systemctl enable ceph-mds@pine02 -[2020-03-11 20:51:08,286][pine02][INFO ] Running command: systemctl start ceph-mds@pine02 -[2020-03-11 20:51:08,479][pine02][INFO ] Running command: systemctl enable ceph.target -[2020-03-11 20:51:13,768][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-11 20:51:13,768][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mds create pine01 -[2020-03-11 20:51:13,768][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-11 20:51:13,768][ceph_deploy.cli][INFO ] username : root -[2020-03-11 20:51:13,768][ceph_deploy.cli][INFO ] verbose : False -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] quiet : False -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] func : -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] mds : [('pine01', 'pine01')] -[2020-03-11 20:51:13,769][ceph_deploy.cli][INFO ] default_release : False -[2020-03-11 20:51:13,769][ceph_deploy.mds][DEBUG ] Deploying mds, cluster ceph hosts pine01:pine01 -[2020-03-11 20:51:23,085][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-11 20:51:23,087][pine01][DEBUG ] detect platform information from remote host -[2020-03-11 20:51:23,495][pine01][DEBUG ] detect machine type -[2020-03-11 20:51:23,571][ceph_deploy.mds][INFO ] Distro info: debian 10.3 buster -[2020-03-11 20:51:23,571][ceph_deploy.mds][DEBUG ] remote host will use systemd -[2020-03-11 20:51:23,573][ceph_deploy.mds][DEBUG ] deploying mds bootstrap to pine01 -[2020-03-11 20:51:23,573][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-11 20:51:24,150][pine01][DEBUG ] create path if it doesn't exist -[2020-03-11 20:51:24,276][pine01][INFO ] Running command: ceph --cluster ceph --name client.bootstrap-mds --keyring /var/lib/ceph/bootstrap-mds/ceph.keyring auth get-or-create mds.pine01 osd allow rwx mds allow mon allow profile mds -o /var/lib/ceph/mds/ceph-pine01/keyring -[2020-03-11 20:51:28,089][pine01][INFO ] Running command: systemctl enable ceph-mds@pine01 -[2020-03-11 20:51:29,088][pine01][INFO ] Running command: systemctl start ceph-mds@pine01 -[2020-03-11 20:51:29,545][pine01][INFO ] Running command: systemctl enable ceph.target -[2020-03-13 19:10:34,252][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sda ebin01 -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] username : root -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] journal : None -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] host : ebin01 -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] filestore : None -[2020-03-13 19:10:34,253][ceph_deploy.cli][INFO ] func : -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] data : /dev/sda -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] block_db : None -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 19:10:34,254][ceph_deploy.cli][INFO ] debug : False -[2020-03-13 19:10:34,255][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sda -[2020-03-13 19:10:37,605][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-13 19:10:37,618][ebin01][DEBUG ] detect platform information from remote host -[2020-03-13 19:10:37,820][ebin01][DEBUG ] detect machine type -[2020-03-13 19:10:37,843][ebin01][DEBUG ] find the location of an executable -[2020-03-13 19:10:37,848][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-13 19:10:37,848][ceph_deploy.osd][DEBUG ] Deploying osd to ebin01 -[2020-03-13 19:10:37,849][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 19:10:37,934][ebin01][DEBUG ] find the location of an executable -[2020-03-13 19:10:37,946][ebin01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-13 19:11:08,840][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:11:08,841][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new fe454719-4b04-414d-a821-07b6ca64602d -[2020-03-13 19:11:08,841][ebin01][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-5dc05fd9-8537-43ef-883c-766d09dfb67f /dev/sda -[2020-03-13 19:11:08,842][ebin01][WARNING] stdout: Wiping dos signature on /dev/sda. -[2020-03-13 19:11:08,845][ebin01][WARNING] stdout: Physical volume "/dev/sda" successfully created. -[2020-03-13 19:11:08,849][ebin01][WARNING] stdout: Volume group "ceph-5dc05fd9-8537-43ef-883c-766d09dfb67f" successfully created -[2020-03-13 19:11:08,850][ebin01][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-fe454719-4b04-414d-a821-07b6ca64602d ceph-5dc05fd9-8537-43ef-883c-766d09dfb67f -[2020-03-13 19:11:08,866][ebin01][WARNING] stdout: Logical volume "osd-block-fe454719-4b04-414d-a821-07b6ca64602d" created. -[2020-03-13 19:11:08,867][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:11:08,868][ebin01][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-4 -[2020-03-13 19:11:08,868][ebin01][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-13 19:11:08,869][ebin01][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-13 19:11:08,870][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-5dc05fd9-8537-43ef-883c-766d09dfb67f/osd-block-fe454719-4b04-414d-a821-07b6ca64602d -[2020-03-13 19:11:08,870][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-2 -[2020-03-13 19:11:08,871][ebin01][WARNING] Running command: /bin/ln -s /dev/ceph-5dc05fd9-8537-43ef-883c-766d09dfb67f/osd-block-fe454719-4b04-414d-a821-07b6ca64602d /var/lib/ceph/osd/ceph-4/block -[2020-03-13 19:11:08,873][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-4/activate.monmap -[2020-03-13 19:11:08,875][ebin01][WARNING] stderr: 2020-03-13 19:10:54.350 7f7b7e61e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-13 19:11:08,876][ebin01][WARNING] 2020-03-13 19:10:54.350 7f7b7e61e0 -1 AuthRegistry(0x7f740814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-13 19:11:08,880][ebin01][WARNING] stderr: got monmap epoch 4 -[2020-03-13 19:11:08,880][ebin01][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-4/keyring --create-keyring --name osd.4 --add-key AQCgzGtesfBxNBAATydWwZgun2+jAo832p9OwA== -[2020-03-13 19:11:08,881][ebin01][WARNING] stdout: creating /var/lib/ceph/osd/ceph-4/keyring -[2020-03-13 19:11:08,883][ebin01][WARNING] added entity osd.4 auth(key=AQCgzGtesfBxNBAATydWwZgun2+jAo832p9OwA==) -[2020-03-13 19:11:08,891][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4/keyring -[2020-03-13 19:11:08,891][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4/ -[2020-03-13 19:11:08,892][ebin01][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 4 --monmap /var/lib/ceph/osd/ceph-4/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-4/ --osd-uuid fe454719-4b04-414d-a821-07b6ca64602d --setuser ceph --setgroup ceph -[2020-03-13 19:11:08,894][ebin01][WARNING] --> ceph-volume lvm prepare successful for: /dev/sda -[2020-03-13 19:11:08,895][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4 -[2020-03-13 19:11:08,897][ebin01][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-5dc05fd9-8537-43ef-883c-766d09dfb67f/osd-block-fe454719-4b04-414d-a821-07b6ca64602d --path /var/lib/ceph/osd/ceph-4 --no-mon-config -[2020-03-13 19:11:08,898][ebin01][WARNING] Running command: /bin/ln -snf /dev/ceph-5dc05fd9-8537-43ef-883c-766d09dfb67f/osd-block-fe454719-4b04-414d-a821-07b6ca64602d /var/lib/ceph/osd/ceph-4/block -[2020-03-13 19:11:08,900][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-4/block -[2020-03-13 19:11:08,901][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-2 -[2020-03-13 19:11:08,903][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-4 -[2020-03-13 19:11:08,904][ebin01][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-4-fe454719-4b04-414d-a821-07b6ca64602d -[2020-03-13 19:11:08,909][ebin01][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-4-fe454719-4b04-414d-a821-07b6ca64602d.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-13 19:11:08,909][ebin01][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@4 -[2020-03-13 19:11:08,913][ebin01][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@4.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-13 19:11:08,914][ebin01][WARNING] Running command: /bin/systemctl start ceph-osd@4 -[2020-03-13 19:11:08,914][ebin01][WARNING] --> ceph-volume lvm activate successful for osd ID: 4 -[2020-03-13 19:11:08,915][ebin01][WARNING] --> ceph-volume lvm create successful for: /dev/sda -[2020-03-13 19:11:14,081][ebin01][INFO ] checking OSD status... -[2020-03-13 19:11:14,082][ebin01][DEBUG ] find the location of an executable -[2020-03-13 19:11:14,095][ebin01][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-13 19:11:15,416][ceph_deploy.osd][DEBUG ] Host ebin01 is now ready for osd use. -[2020-03-13 19:11:26,241][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 19:11:26,242][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sdb ebin01 -[2020-03-13 19:11:26,289][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 19:11:26,289][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 19:11:26,289][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-13 19:11:26,290][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 19:11:26,290][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 19:11:26,290][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-13 19:11:26,290][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-13 19:11:26,290][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 19:11:26,291][ceph_deploy.cli][INFO ] username : root -[2020-03-13 19:11:26,291][ceph_deploy.cli][INFO ] journal : None -[2020-03-13 19:11:26,291][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-13 19:11:26,291][ceph_deploy.cli][INFO ] host : ebin01 -[2020-03-13 19:11:26,291][ceph_deploy.cli][INFO ] filestore : None -[2020-03-13 19:11:26,291][ceph_deploy.cli][INFO ] func : -[2020-03-13 19:11:26,292][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 19:11:26,292][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-13 19:11:26,292][ceph_deploy.cli][INFO ] data : /dev/sdb -[2020-03-13 19:11:26,292][ceph_deploy.cli][INFO ] block_db : None -[2020-03-13 19:11:26,292][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-13 19:11:26,292][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-13 19:11:26,293][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-13 19:11:26,293][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 19:11:26,293][ceph_deploy.cli][INFO ] debug : False -[2020-03-13 19:11:26,295][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdb -[2020-03-13 19:11:32,575][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-13 19:11:32,577][ebin01][DEBUG ] detect platform information from remote host -[2020-03-13 19:11:32,994][ebin01][DEBUG ] detect machine type -[2020-03-13 19:11:33,032][ebin01][DEBUG ] find the location of an executable -[2020-03-13 19:11:33,054][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-13 19:11:33,054][ceph_deploy.osd][DEBUG ] Deploying osd to ebin01 -[2020-03-13 19:11:33,055][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 19:11:33,508][ebin01][DEBUG ] find the location of an executable -[2020-03-13 19:11:33,562][ebin01][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdb -[2020-03-13 19:12:32,294][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:12:32,298][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 5f8d00a1-bc65-4202-bfc5-c0eab2aba52b -[2020-03-13 19:12:32,314][ebin01][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-f1986b95-1069-454c-ab67-e588f43407a6 /dev/sdb -[2020-03-13 19:12:32,314][ebin01][WARNING] stdout: Physical volume "/dev/sdb" successfully created. -[2020-03-13 19:12:32,319][ebin01][WARNING] stdout: Volume group "ceph-f1986b95-1069-454c-ab67-e588f43407a6" successfully created -[2020-03-13 19:12:32,335][ebin01][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b ceph-f1986b95-1069-454c-ab67-e588f43407a6 -[2020-03-13 19:12:32,343][ebin01][WARNING] stdout: Logical volume "osd-block-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b" created. -[2020-03-13 19:12:32,351][ebin01][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:12:32,355][ebin01][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-5 -[2020-03-13 19:12:32,359][ebin01][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-13 19:12:32,375][ebin01][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-13 19:12:32,377][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-f1986b95-1069-454c-ab67-e588f43407a6/osd-block-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b -[2020-03-13 19:12:32,384][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-3 -[2020-03-13 19:12:32,402][ebin01][WARNING] Running command: /bin/ln -s /dev/ceph-f1986b95-1069-454c-ab67-e588f43407a6/osd-block-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b /var/lib/ceph/osd/ceph-5/block -[2020-03-13 19:12:32,418][ebin01][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-5/activate.monmap -[2020-03-13 19:12:32,426][ebin01][WARNING] stderr: 2020-03-13 19:12:10.073 7f916761e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-13 19:12:32,428][ebin01][WARNING] stderr: -[2020-03-13 19:12:32,443][ebin01][WARNING] stderr: 2020-03-13 19:12:10.073 7f916761e0 -1 AuthRegistry(0x7f8c0814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-13 19:12:32,444][ebin01][WARNING] stderr: got monmap epoch 4 -[2020-03-13 19:12:32,444][ebin01][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-5/keyring --create-keyring --name osd.5 --add-key AQDfzGte6I5MAxAAlFVdK/KhPHmFopRKQxm89g== -[2020-03-13 19:12:32,444][ebin01][WARNING] stdout: creating /var/lib/ceph/osd/ceph-5/keyring -[2020-03-13 19:12:32,445][ebin01][WARNING] added entity osd.5 auth(key=AQDfzGte6I5MAxAAlFVdK/KhPHmFopRKQxm89g==) -[2020-03-13 19:12:32,453][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5/keyring -[2020-03-13 19:12:32,453][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5/ -[2020-03-13 19:12:32,469][ebin01][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 5 --monmap /var/lib/ceph/osd/ceph-5/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-5/ --osd-uuid 5f8d00a1-bc65-4202-bfc5-c0eab2aba52b --setuser ceph --setgroup ceph -[2020-03-13 19:12:32,469][ebin01][WARNING] --> ceph-volume lvm prepare successful for: /dev/sdb -[2020-03-13 19:12:32,470][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 -[2020-03-13 19:12:32,473][ebin01][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-f1986b95-1069-454c-ab67-e588f43407a6/osd-block-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b --path /var/lib/ceph/osd/ceph-5 --no-mon-config -[2020-03-13 19:12:32,489][ebin01][WARNING] Running command: /bin/ln -snf /dev/ceph-f1986b95-1069-454c-ab67-e588f43407a6/osd-block-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b /var/lib/ceph/osd/ceph-5/block -[2020-03-13 19:12:32,497][ebin01][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-5/block -[2020-03-13 19:12:32,498][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-3 -[2020-03-13 19:12:32,506][ebin01][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 -[2020-03-13 19:12:32,510][ebin01][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-5-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b -[2020-03-13 19:12:32,525][ebin01][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-5-5f8d00a1-bc65-4202-bfc5-c0eab2aba52b.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-13 19:12:32,526][ebin01][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@5 -[2020-03-13 19:12:32,534][ebin01][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@5.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-13 19:12:32,534][ebin01][WARNING] Running command: /bin/systemctl start ceph-osd@5 -[2020-03-13 19:12:32,538][ebin01][WARNING] --> ceph-volume lvm activate successful for osd ID: 5 -[2020-03-13 19:12:32,538][ebin01][WARNING] --> ceph-volume lvm create successful for: /dev/sdb -[2020-03-13 19:12:37,657][ebin01][INFO ] checking OSD status... -[2020-03-13 19:12:37,658][ebin01][DEBUG ] find the location of an executable -[2020-03-13 19:12:37,695][ebin01][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-13 19:12:41,977][ceph_deploy.osd][DEBUG ] Host ebin01 is now ready for osd use. -[2020-03-13 19:12:54,371][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sda ebin02 -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] username : root -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] journal : None -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] host : ebin02 -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] filestore : None -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] func : -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] data : /dev/sda -[2020-03-13 19:12:54,372][ceph_deploy.cli][INFO ] block_db : None -[2020-03-13 19:12:54,373][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-13 19:12:54,373][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-13 19:12:54,373][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-13 19:12:54,373][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 19:12:54,373][ceph_deploy.cli][INFO ] debug : False -[2020-03-13 19:12:54,373][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sda -[2020-03-13 19:12:57,729][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-13 19:12:57,731][ebin02][DEBUG ] detect platform information from remote host -[2020-03-13 19:12:57,962][ebin02][DEBUG ] detect machine type -[2020-03-13 19:12:57,989][ebin02][DEBUG ] find the location of an executable -[2020-03-13 19:12:57,996][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-13 19:12:57,997][ceph_deploy.osd][DEBUG ] Deploying osd to ebin02 -[2020-03-13 19:12:57,999][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 19:12:58,042][ebin02][DEBUG ] find the location of an executable -[2020-03-13 19:12:58,058][ebin02][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda -[2020-03-13 19:13:29,844][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:13:29,845][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 1e329b82-74f3-42b9-bc1e-8b115324d402 -[2020-03-13 19:13:29,845][ebin02][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-ed4c22ce-df90-4f83-861f-3e071de52ccf /dev/sda -[2020-03-13 19:13:29,845][ebin02][WARNING] stdout: Wiping dos signature on /dev/sda. -[2020-03-13 19:13:29,845][ebin02][WARNING] stdout: Physical volume "/dev/sda" successfully created. -[2020-03-13 19:13:29,861][ebin02][WARNING] stdout: Volume group "ceph-ed4c22ce-df90-4f83-861f-3e071de52ccf" successfully created -[2020-03-13 19:13:29,862][ebin02][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-1e329b82-74f3-42b9-bc1e-8b115324d402 ceph-ed4c22ce-df90-4f83-861f-3e071de52ccf -[2020-03-13 19:13:29,863][ebin02][WARNING] stdout: Logical volume "osd-block-1e329b82-74f3-42b9-bc1e-8b115324d402" created. -[2020-03-13 19:13:29,863][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:13:29,863][ebin02][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-6 -[2020-03-13 19:13:29,863][ebin02][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-13 19:13:29,864][ebin02][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-13 19:13:29,864][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-ed4c22ce-df90-4f83-861f-3e071de52ccf/osd-block-1e329b82-74f3-42b9-bc1e-8b115324d402 -[2020-03-13 19:13:29,864][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-2 -[2020-03-13 19:13:29,864][ebin02][WARNING] Running command: /bin/ln -s /dev/ceph-ed4c22ce-df90-4f83-861f-3e071de52ccf/osd-block-1e329b82-74f3-42b9-bc1e-8b115324d402 /var/lib/ceph/osd/ceph-6/block -[2020-03-13 19:13:29,865][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-6/activate.monmap -[2020-03-13 19:13:29,881][ebin02][WARNING] stderr: 2020-03-13 19:13:13.918 7f78b631e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-13 19:13:29,881][ebin02][WARNING] stderr: 2020-03-13 19:13:13.918 7f78b631e0 -1 AuthRegistry(0x7f740814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-13 19:13:29,883][ebin02][WARNING] stderr: got monmap epoch 4 -[2020-03-13 19:13:29,883][ebin02][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-6/keyring --create-keyring --name osd.6 --add-key AQAuzWte6sTKLRAARTnzg5C6ogvte9QnmYmqsg== -[2020-03-13 19:13:29,884][ebin02][WARNING] stdout: creating /var/lib/ceph/osd/ceph-6/keyring -[2020-03-13 19:13:29,884][ebin02][WARNING] added entity osd.6 auth(key=AQAuzWte6sTKLRAARTnzg5C6ogvte9QnmYmqsg==) -[2020-03-13 19:13:29,884][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6/keyring -[2020-03-13 19:13:29,885][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6/ -[2020-03-13 19:13:29,885][ebin02][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 6 --monmap /var/lib/ceph/osd/ceph-6/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-6/ --osd-uuid 1e329b82-74f3-42b9-bc1e-8b115324d402 --setuser ceph --setgroup ceph -[2020-03-13 19:13:29,885][ebin02][WARNING] --> ceph-volume lvm prepare successful for: /dev/sda -[2020-03-13 19:13:29,901][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 -[2020-03-13 19:13:29,901][ebin02][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-ed4c22ce-df90-4f83-861f-3e071de52ccf/osd-block-1e329b82-74f3-42b9-bc1e-8b115324d402 --path /var/lib/ceph/osd/ceph-6 --no-mon-config -[2020-03-13 19:13:29,902][ebin02][WARNING] Running command: /bin/ln -snf /dev/ceph-ed4c22ce-df90-4f83-861f-3e071de52ccf/osd-block-1e329b82-74f3-42b9-bc1e-8b115324d402 /var/lib/ceph/osd/ceph-6/block -[2020-03-13 19:13:29,902][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-6/block -[2020-03-13 19:13:29,902][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-2 -[2020-03-13 19:13:29,903][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 -[2020-03-13 19:13:29,910][ebin02][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-6-1e329b82-74f3-42b9-bc1e-8b115324d402 -[2020-03-13 19:13:29,911][ebin02][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-6-1e329b82-74f3-42b9-bc1e-8b115324d402.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-13 19:13:29,911][ebin02][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@6 -[2020-03-13 19:13:29,911][ebin02][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@6.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-13 19:13:29,912][ebin02][WARNING] Running command: /bin/systemctl start ceph-osd@6 -[2020-03-13 19:13:29,928][ebin02][WARNING] --> ceph-volume lvm activate successful for osd ID: 6 -[2020-03-13 19:13:29,928][ebin02][WARNING] --> ceph-volume lvm create successful for: /dev/sda -[2020-03-13 19:13:34,997][ebin02][INFO ] checking OSD status... -[2020-03-13 19:13:34,998][ebin02][DEBUG ] find the location of an executable -[2020-03-13 19:13:35,032][ebin02][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-13 19:13:36,554][ceph_deploy.osd][DEBUG ] Host ebin02 is now ready for osd use. -[2020-03-13 19:13:54,082][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root osd create --data /dev/sdb ebin02 -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] bluestore : None -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] fs_type : xfs -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] block_wal : None -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] username : root -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] journal : None -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] host : ebin02 -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] filestore : None -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] func : -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 19:13:54,083][ceph_deploy.cli][INFO ] zap_disk : False -[2020-03-13 19:13:54,084][ceph_deploy.cli][INFO ] data : /dev/sdb -[2020-03-13 19:13:54,084][ceph_deploy.cli][INFO ] block_db : None -[2020-03-13 19:13:54,084][ceph_deploy.cli][INFO ] dmcrypt : False -[2020-03-13 19:13:54,084][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-13 19:13:54,084][ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys -[2020-03-13 19:13:54,084][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 19:13:54,084][ceph_deploy.cli][INFO ] debug : False -[2020-03-13 19:13:54,084][ceph_deploy.osd][DEBUG ] Creating OSD on cluster ceph with data device /dev/sdb -[2020-03-13 19:13:57,070][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-13 19:13:57,072][ebin02][DEBUG ] detect platform information from remote host -[2020-03-13 19:13:57,299][ebin02][DEBUG ] detect machine type -[2020-03-13 19:13:57,327][ebin02][DEBUG ] find the location of an executable -[2020-03-13 19:13:57,332][ceph_deploy.osd][INFO ] Distro info: debian 10.3 buster -[2020-03-13 19:13:57,332][ceph_deploy.osd][DEBUG ] Deploying osd to ebin02 -[2020-03-13 19:13:57,334][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 19:13:57,734][ebin02][DEBUG ] find the location of an executable -[2020-03-13 19:13:57,747][ebin02][INFO ] Running command: /usr/sbin/ceph-volume --cluster ceph lvm create --bluestore --data /dev/sdb -[2020-03-13 19:14:28,543][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:14:28,543][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 24c2823e-d63b-4b1f-9c83-76cae58bbb38 -[2020-03-13 19:14:28,544][ebin02][WARNING] Running command: /sbin/vgcreate -s 1G --force --yes ceph-4e945599-23d6-4222-bd1f-017a4a869912 /dev/sdb -[2020-03-13 19:14:28,544][ebin02][WARNING] stdout: Physical volume "/dev/sdb" successfully created. -[2020-03-13 19:14:28,544][ebin02][WARNING] stdout: Volume group "ceph-4e945599-23d6-4222-bd1f-017a4a869912" successfully created -[2020-03-13 19:14:28,544][ebin02][WARNING] Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-24c2823e-d63b-4b1f-9c83-76cae58bbb38 ceph-4e945599-23d6-4222-bd1f-017a4a869912 -[2020-03-13 19:14:28,545][ebin02][WARNING] stdout: Logical volume "osd-block-24c2823e-d63b-4b1f-9c83-76cae58bbb38" created. -[2020-03-13 19:14:28,545][ebin02][WARNING] Running command: /usr/bin/ceph-authtool --gen-print-key -[2020-03-13 19:14:28,545][ebin02][WARNING] Running command: /bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-7 -[2020-03-13 19:14:28,546][ebin02][WARNING] --> Absolute path not found for executable: selinuxenabled -[2020-03-13 19:14:28,546][ebin02][WARNING] --> Ensure $PATH environment variable contains common executable locations -[2020-03-13 19:14:28,546][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /dev/ceph-4e945599-23d6-4222-bd1f-017a4a869912/osd-block-24c2823e-d63b-4b1f-9c83-76cae58bbb38 -[2020-03-13 19:14:28,546][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-3 -[2020-03-13 19:14:28,546][ebin02][WARNING] Running command: /bin/ln -s /dev/ceph-4e945599-23d6-4222-bd1f-017a4a869912/osd-block-24c2823e-d63b-4b1f-9c83-76cae58bbb38 /var/lib/ceph/osd/ceph-7/block -[2020-03-13 19:14:28,547][ebin02][WARNING] Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-7/activate.monmap -[2020-03-13 19:14:28,547][ebin02][WARNING] stderr: 2020-03-13 19:14:13.452 7f7e1ac1e0 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,: (2) No such file or directory -[2020-03-13 19:14:28,549][ebin02][WARNING] stderr: 2020-03-13 19:14:13.452 7f7e1ac1e0 -1 AuthRegistry(0x7f780814b8) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,, disabling cephx -[2020-03-13 19:14:28,550][ebin02][WARNING] stderr: got monmap epoch 4 -[2020-03-13 19:14:28,552][ebin02][WARNING] Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-7/keyring --create-keyring --name osd.7 --add-key AQBpzWteiaeEMBAAakF8uj1EEE8fC21M3BvzMA== -[2020-03-13 19:14:28,553][ebin02][WARNING] stdout: creating /var/lib/ceph/osd/ceph-7/keyring -[2020-03-13 19:14:28,553][ebin02][WARNING] added entity osd.7 auth(key=AQBpzWteiaeEMBAAakF8uj1EEE8fC21M3BvzMA==) -[2020-03-13 19:14:28,554][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7/keyring -[2020-03-13 19:14:28,555][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7/ -[2020-03-13 19:14:28,556][ebin02][WARNING] Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 7 --monmap /var/lib/ceph/osd/ceph-7/activate.monmap --keyfile - --osd-data /var/lib/ceph/osd/ceph-7/ --osd-uuid 24c2823e-d63b-4b1f-9c83-76cae58bbb38 --setuser ceph --setgroup ceph -[2020-03-13 19:14:28,557][ebin02][WARNING] --> ceph-volume lvm prepare successful for: /dev/sdb -[2020-03-13 19:14:28,557][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 -[2020-03-13 19:14:28,564][ebin02][WARNING] Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/ceph-4e945599-23d6-4222-bd1f-017a4a869912/osd-block-24c2823e-d63b-4b1f-9c83-76cae58bbb38 --path /var/lib/ceph/osd/ceph-7 --no-mon-config -[2020-03-13 19:14:28,565][ebin02][WARNING] Running command: /bin/ln -snf /dev/ceph-4e945599-23d6-4222-bd1f-017a4a869912/osd-block-24c2823e-d63b-4b1f-9c83-76cae58bbb38 /var/lib/ceph/osd/ceph-7/block -[2020-03-13 19:14:28,565][ebin02][WARNING] Running command: /bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-7/block -[2020-03-13 19:14:28,566][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /dev/dm-3 -[2020-03-13 19:14:28,566][ebin02][WARNING] Running command: /bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 -[2020-03-13 19:14:28,567][ebin02][WARNING] Running command: /bin/systemctl enable ceph-volume@lvm-7-24c2823e-d63b-4b1f-9c83-76cae58bbb38 -[2020-03-13 19:14:28,583][ebin02][WARNING] stderr: Created symlink /etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-7-24c2823e-d63b-4b1f-9c83-76cae58bbb38.service → /lib/systemd/system/ceph-volume@.service. -[2020-03-13 19:14:28,583][ebin02][WARNING] Running command: /bin/systemctl enable --runtime ceph-osd@7 -[2020-03-13 19:14:28,584][ebin02][WARNING] stderr: Created symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@7.service → /lib/systemd/system/ceph-osd@.service. -[2020-03-13 19:14:28,584][ebin02][WARNING] Running command: /bin/systemctl start ceph-osd@7 -[2020-03-13 19:14:28,584][ebin02][WARNING] --> ceph-volume lvm activate successful for osd ID: 7 -[2020-03-13 19:14:28,588][ebin02][WARNING] --> ceph-volume lvm create successful for: /dev/sdb -[2020-03-13 19:14:33,757][ebin02][INFO ] checking OSD status... -[2020-03-13 19:14:33,778][ebin02][DEBUG ] find the location of an executable -[2020-03-13 19:14:33,802][ebin02][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json -[2020-03-13 19:14:35,323][ceph_deploy.osd][DEBUG ] Host ebin02 is now ready for osd use. -[2020-03-13 21:51:47,721][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 21:51:47,728][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny tumor pine03 -[2020-03-13 21:51:47,728][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 21:51:47,728][ceph_deploy.cli][INFO ] username : root -[2020-03-13 21:51:47,728][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 21:51:47,728][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-13 21:51:47,728][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-13 21:51:47,729][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 21:51:47,729][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 21:51:47,729][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 21:51:47,729][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny', 'tumor', 'pine03'] -[2020-03-13 21:51:47,729][ceph_deploy.cli][INFO ] func : -[2020-03-13 21:51:47,729][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 21:51:47,729][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 21:51:47,729][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-13 21:51:49,243][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-13 21:51:49,246][riot01][DEBUG ] detect platform information from remote host -[2020-03-13 21:51:49,420][riot01][DEBUG ] detect machine type -[2020-03-13 21:51:49,456][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 21:51:49,475][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-13 21:51:58,175][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-13 21:51:58,178][pine01][DEBUG ] detect platform information from remote host -[2020-03-13 21:51:59,230][pine01][DEBUG ] detect machine type -[2020-03-13 21:51:59,430][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 21:51:59,439][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-13 21:52:01,198][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-13 21:52:01,200][pine02][DEBUG ] detect platform information from remote host -[2020-03-13 21:52:01,357][pine02][DEBUG ] detect machine type -[2020-03-13 21:52:01,376][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 21:52:01,393][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-13 21:52:10,272][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-13 21:52:10,274][ebin01][DEBUG ] detect platform information from remote host -[2020-03-13 21:52:10,950][ebin01][DEBUG ] detect machine type -[2020-03-13 21:52:11,008][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 21:52:11,049][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-13 21:52:17,591][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-13 21:52:17,593][ebin02][DEBUG ] detect platform information from remote host -[2020-03-13 21:52:18,045][ebin02][DEBUG ] detect machine type -[2020-03-13 21:52:18,068][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 21:52:18,096][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-13 21:52:18,636][lenny][DEBUG ] connected to host: root@lenny -[2020-03-13 21:52:18,636][lenny][DEBUG ] detect platform information from remote host -[2020-03-13 21:52:18,653][lenny][DEBUG ] detect machine type -[2020-03-13 21:52:18,657][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 21:52:18,659][ceph_deploy.config][DEBUG ] Pushing config to tumor -[2020-03-13 21:52:19,641][tumor][DEBUG ] connected to host: root@tumor -[2020-03-13 21:52:19,642][tumor][DEBUG ] detect platform information from remote host -[2020-03-13 21:52:19,813][tumor][DEBUG ] detect machine type -[2020-03-13 21:52:19,833][tumor][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 21:52:19,849][ceph_deploy.config][DEBUG ] Pushing config to pine03 -[2020-03-13 21:52:30,572][pine03][DEBUG ] connected to host: root@pine03 -[2020-03-13 21:52:30,574][pine03][DEBUG ] detect platform information from remote host -[2020-03-13 21:52:31,298][pine03][DEBUG ] detect machine type -[2020-03-13 21:52:31,372][pine03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:15:59,880][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root mon create pine03 -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] username : root -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] overwrite_conf : False -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] subcommand : create -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] mon : ['pine03'] -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] func : -[2020-03-13 22:15:59,880][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 22:15:59,881][ceph_deploy.cli][INFO ] keyrings : None -[2020-03-13 22:15:59,881][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 22:15:59,891][ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts pine03 -[2020-03-13 22:15:59,891][ceph_deploy.mon][DEBUG ] detecting platform for host pine03 ... -[2020-03-13 22:16:07,142][pine03][DEBUG ] connected to host: root@pine03 -[2020-03-13 22:16:07,143][pine03][DEBUG ] detect platform information from remote host -[2020-03-13 22:16:07,805][pine03][DEBUG ] detect machine type -[2020-03-13 22:16:07,940][pine03][DEBUG ] find the location of an executable -[2020-03-13 22:16:08,003][ceph_deploy.mon][INFO ] distro info: debian 10.3 buster -[2020-03-13 22:16:08,004][pine03][DEBUG ] determining if provided host has same hostname in remote -[2020-03-13 22:16:08,004][pine03][DEBUG ] get remote short hostname -[2020-03-13 22:16:08,069][pine03][DEBUG ] deploying mon to pine03 -[2020-03-13 22:16:08,070][pine03][DEBUG ] get remote short hostname -[2020-03-13 22:16:08,074][pine03][DEBUG ] remote hostname: pine03 -[2020-03-13 22:16:08,146][pine03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:16:08,212][pine03][DEBUG ] create the mon path if it does not exist -[2020-03-13 22:16:08,290][pine03][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-pine03/done -[2020-03-13 22:16:08,294][pine03][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-pine03/done -[2020-03-13 22:16:08,352][pine03][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-pine03.mon.keyring -[2020-03-13 22:16:08,353][pine03][DEBUG ] create the monitor keyring file -[2020-03-13 22:16:08,459][pine03][INFO ] Running command: ceph-mon --cluster ceph --mkfs -i pine03 --keyring /var/lib/ceph/tmp/ceph-pine03.mon.keyring --setuser 64045 --setgroup 64045 -[2020-03-13 22:16:20,798][pine03][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-pine03.mon.keyring -[2020-03-13 22:16:20,819][pine03][DEBUG ] create a done file to avoid re-doing the mon deployment -[2020-03-13 22:16:20,823][pine03][DEBUG ] create the init path if it does not exist -[2020-03-13 22:16:20,890][pine03][INFO ] Running command: systemctl enable ceph.target -[2020-03-13 22:16:22,207][pine03][INFO ] Running command: systemctl enable ceph-mon@pine03 -[2020-03-13 22:16:22,480][pine03][WARNING] Created symlink /etc/systemd/system/ceph-mon.target.wants/ceph-mon@pine03.service → /lib/systemd/system/ceph-mon@.service. -[2020-03-13 22:16:23,615][pine03][INFO ] Running command: systemctl start ceph-mon@pine03 -[2020-03-13 22:16:26,078][pine03][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine03.asok mon_status -[2020-03-13 22:16:30,208][pine03][DEBUG ] ******************************************************************************** -[2020-03-13 22:16:30,208][pine03][DEBUG ] status for monitor: mon.pine03 -[2020-03-13 22:16:30,210][pine03][DEBUG ] { -[2020-03-13 22:16:30,210][pine03][DEBUG ] "election_epoch": 0, -[2020-03-13 22:16:30,210][pine03][DEBUG ] "extra_probe_peers": [ -[2020-03-13 22:16:30,210][pine03][DEBUG ] { -[2020-03-13 22:16:30,211][pine03][DEBUG ] "addrvec": [ -[2020-03-13 22:16:30,211][pine03][DEBUG ] { -[2020-03-13 22:16:30,211][pine03][DEBUG ] "addr": "192.168.10.160:3300", -[2020-03-13 22:16:30,211][pine03][DEBUG ] "nonce": 0, -[2020-03-13 22:16:30,211][pine03][DEBUG ] "type": "v2" -[2020-03-13 22:16:30,212][pine03][DEBUG ] }, -[2020-03-13 22:16:30,212][pine03][DEBUG ] { -[2020-03-13 22:16:30,212][pine03][DEBUG ] "addr": "192.168.10.160:6789", -[2020-03-13 22:16:30,212][pine03][DEBUG ] "nonce": 0, -[2020-03-13 22:16:30,212][pine03][DEBUG ] "type": "v1" -[2020-03-13 22:16:30,212][pine03][DEBUG ] } -[2020-03-13 22:16:30,213][pine03][DEBUG ] ] -[2020-03-13 22:16:30,213][pine03][DEBUG ] }, -[2020-03-13 22:16:30,213][pine03][DEBUG ] { -[2020-03-13 22:16:30,213][pine03][DEBUG ] "addrvec": [ -[2020-03-13 22:16:30,214][pine03][DEBUG ] { -[2020-03-13 22:16:30,214][pine03][DEBUG ] "addr": "192.168.10.161:3300", -[2020-03-13 22:16:30,214][pine03][DEBUG ] "nonce": 0, -[2020-03-13 22:16:30,214][pine03][DEBUG ] "type": "v2" -[2020-03-13 22:16:30,214][pine03][DEBUG ] }, -[2020-03-13 22:16:30,214][pine03][DEBUG ] { -[2020-03-13 22:16:30,215][pine03][DEBUG ] "addr": "192.168.10.161:6789", -[2020-03-13 22:16:30,215][pine03][DEBUG ] "nonce": 0, -[2020-03-13 22:16:30,215][pine03][DEBUG ] "type": "v1" -[2020-03-13 22:16:30,215][pine03][DEBUG ] } -[2020-03-13 22:16:30,215][pine03][DEBUG ] ] -[2020-03-13 22:16:30,216][pine03][DEBUG ] } -[2020-03-13 22:16:30,216][pine03][DEBUG ] ], -[2020-03-13 22:16:30,216][pine03][DEBUG ] "feature_map": { -[2020-03-13 22:16:30,216][pine03][DEBUG ] "mon": [ -[2020-03-13 22:16:30,217][pine03][DEBUG ] { -[2020-03-13 22:16:30,217][pine03][DEBUG ] "features": "0x3ffddff8ffacffff", -[2020-03-13 22:16:30,217][pine03][DEBUG ] "num": 1, -[2020-03-13 22:16:30,217][pine03][DEBUG ] "release": "luminous" -[2020-03-13 22:16:30,217][pine03][DEBUG ] } -[2020-03-13 22:16:30,218][pine03][DEBUG ] ] -[2020-03-13 22:16:30,218][pine03][DEBUG ] }, -[2020-03-13 22:16:30,218][pine03][DEBUG ] "features": { -[2020-03-13 22:16:30,218][pine03][DEBUG ] "quorum_con": "0", -[2020-03-13 22:16:30,218][pine03][DEBUG ] "quorum_mon": [], -[2020-03-13 22:16:30,219][pine03][DEBUG ] "required_con": "0", -[2020-03-13 22:16:30,219][pine03][DEBUG ] "required_mon": [] -[2020-03-13 22:16:30,219][pine03][DEBUG ] }, -[2020-03-13 22:16:30,220][pine03][DEBUG ] "monmap": { -[2020-03-13 22:16:30,220][pine03][DEBUG ] "created": "2020-03-13 22:16:14.498641", -[2020-03-13 22:16:30,220][pine03][DEBUG ] "epoch": 0, -[2020-03-13 22:16:30,220][pine03][DEBUG ] "features": { -[2020-03-13 22:16:30,221][pine03][DEBUG ] "optional": [], -[2020-03-13 22:16:30,221][pine03][DEBUG ] "persistent": [] -[2020-03-13 22:16:30,221][pine03][DEBUG ] }, -[2020-03-13 22:16:30,221][pine03][DEBUG ] "fsid": "29ef4020-303a-4b2e-aa24-a1e20e5ba21c", -[2020-03-13 22:16:30,221][pine03][DEBUG ] "min_mon_release": 0, -[2020-03-13 22:16:30,222][pine03][DEBUG ] "min_mon_release_name": "unknown", -[2020-03-13 22:16:30,222][pine03][DEBUG ] "modified": "2020-03-13 22:16:14.498641", -[2020-03-13 22:16:30,222][pine03][DEBUG ] "mons": [ -[2020-03-13 22:16:30,222][pine03][DEBUG ] { -[2020-03-13 22:16:30,223][pine03][DEBUG ] "addr": "0.0.0.0:0/1", -[2020-03-13 22:16:30,223][pine03][DEBUG ] "name": "pine01", -[2020-03-13 22:16:30,223][pine03][DEBUG ] "public_addr": "0.0.0.0:0/1", -[2020-03-13 22:16:30,223][pine03][DEBUG ] "public_addrs": { -[2020-03-13 22:16:30,224][pine03][DEBUG ] "addrvec": [ -[2020-03-13 22:16:30,224][pine03][DEBUG ] { -[2020-03-13 22:16:30,224][pine03][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-13 22:16:30,224][pine03][DEBUG ] "nonce": 1, -[2020-03-13 22:16:30,224][pine03][DEBUG ] "type": "v1" -[2020-03-13 22:16:30,224][pine03][DEBUG ] } -[2020-03-13 22:16:30,225][pine03][DEBUG ] ] -[2020-03-13 22:16:30,225][pine03][DEBUG ] }, -[2020-03-13 22:16:30,225][pine03][DEBUG ] "rank": 0 -[2020-03-13 22:16:30,225][pine03][DEBUG ] }, -[2020-03-13 22:16:30,225][pine03][DEBUG ] { -[2020-03-13 22:16:30,225][pine03][DEBUG ] "addr": "0.0.0.0:0/2", -[2020-03-13 22:16:30,225][pine03][DEBUG ] "name": "pine02", -[2020-03-13 22:16:30,225][pine03][DEBUG ] "public_addr": "0.0.0.0:0/2", -[2020-03-13 22:16:30,225][pine03][DEBUG ] "public_addrs": { -[2020-03-13 22:16:30,226][pine03][DEBUG ] "addrvec": [ -[2020-03-13 22:16:30,226][pine03][DEBUG ] { -[2020-03-13 22:16:30,226][pine03][DEBUG ] "addr": "0.0.0.0:0", -[2020-03-13 22:16:30,226][pine03][DEBUG ] "nonce": 2, -[2020-03-13 22:16:30,226][pine03][DEBUG ] "type": "v1" -[2020-03-13 22:16:30,226][pine03][DEBUG ] } -[2020-03-13 22:16:30,226][pine03][DEBUG ] ] -[2020-03-13 22:16:30,226][pine03][DEBUG ] }, -[2020-03-13 22:16:30,226][pine03][DEBUG ] "rank": 1 -[2020-03-13 22:16:30,227][pine03][DEBUG ] } -[2020-03-13 22:16:30,227][pine03][DEBUG ] ] -[2020-03-13 22:16:30,227][pine03][DEBUG ] }, -[2020-03-13 22:16:30,227][pine03][DEBUG ] "name": "pine03", -[2020-03-13 22:16:30,227][pine03][DEBUG ] "outside_quorum": [], -[2020-03-13 22:16:30,227][pine03][DEBUG ] "quorum": [], -[2020-03-13 22:16:30,227][pine03][DEBUG ] "rank": -1, -[2020-03-13 22:16:30,227][pine03][DEBUG ] "state": "probing", -[2020-03-13 22:16:30,227][pine03][DEBUG ] "sync_provider": [] -[2020-03-13 22:16:30,228][pine03][DEBUG ] } -[2020-03-13 22:16:30,228][pine03][DEBUG ] ******************************************************************************** -[2020-03-13 22:16:30,228][pine03][INFO ] monitor: mon.pine03 is currently at the state of probing -[2020-03-13 22:16:30,315][pine03][INFO ] Running command: ceph --cluster=ceph --admin-daemon /var/run/ceph/ceph-mon.pine03.asok mon_status -[2020-03-13 22:16:31,286][pine03][WARNING] pine03 is not defined in `mon initial members` -[2020-03-13 22:16:31,286][pine03][WARNING] monitor pine03 does not exist in monmap -[2020-03-13 22:18:11,967][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 22:18:11,968][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny tumor pine03 -[2020-03-13 22:18:11,977][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 22:18:11,977][ceph_deploy.cli][INFO ] username : root -[2020-03-13 22:18:11,977][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 22:18:11,977][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny', 'tumor', 'pine03'] -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] func : -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 22:18:11,978][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 22:18:11,978][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-13 22:18:13,472][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-13 22:18:13,474][riot01][DEBUG ] detect platform information from remote host -[2020-03-13 22:18:13,642][riot01][DEBUG ] detect machine type -[2020-03-13 22:18:13,676][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:18:13,698][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-13 22:20:24,023][ceph_deploy.config][ERROR ] connecting to host: root@pine01 resulted in errors: HostNotFound root@pine01 -[2020-03-13 22:20:24,024][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-13 22:20:25,740][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-13 22:20:25,742][pine02][DEBUG ] detect platform information from remote host -[2020-03-13 22:20:25,898][pine02][DEBUG ] detect machine type -[2020-03-13 22:20:25,917][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:20:25,932][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-13 22:20:28,312][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-13 22:20:28,314][ebin01][DEBUG ] detect platform information from remote host -[2020-03-13 22:20:28,526][ebin01][DEBUG ] detect machine type -[2020-03-13 22:20:28,549][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:20:28,561][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-13 22:20:30,716][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-13 22:20:30,718][ebin02][DEBUG ] detect platform information from remote host -[2020-03-13 22:20:30,884][ebin02][DEBUG ] detect machine type -[2020-03-13 22:20:30,925][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:20:30,937][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-13 22:20:31,677][lenny][DEBUG ] connected to host: root@lenny -[2020-03-13 22:20:31,678][lenny][DEBUG ] detect platform information from remote host -[2020-03-13 22:20:31,706][lenny][DEBUG ] detect machine type -[2020-03-13 22:20:31,710][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:20:31,712][ceph_deploy.config][DEBUG ] Pushing config to tumor -[2020-03-13 22:20:32,702][tumor][DEBUG ] connected to host: root@tumor -[2020-03-13 22:20:32,705][tumor][DEBUG ] detect platform information from remote host -[2020-03-13 22:20:32,848][tumor][DEBUG ] detect machine type -[2020-03-13 22:20:32,868][tumor][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:20:32,880][ceph_deploy.config][DEBUG ] Pushing config to pine03 -[2020-03-13 22:20:39,513][pine03][DEBUG ] connected to host: root@pine03 -[2020-03-13 22:20:39,514][pine03][DEBUG ] detect platform information from remote host -[2020-03-13 22:20:39,680][pine03][DEBUG ] detect machine type -[2020-03-13 22:20:39,812][pine03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:20:42,354][ceph_deploy][ERROR ] GenericError: Failed to config 1 hosts - -[2020-03-13 22:46:13,300][ceph_deploy.conf][DEBUG ] found configuration file at: /home/do/.cephdeploy.conf -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] Invoked (2.0.1): /usr/bin/ceph-deploy --username root --overwrite-conf config push riot01 pine01 pine02 ebin01 ebin02 lenny pine03 -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] ceph-deploy options: -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] username : root -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] verbose : False -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] overwrite_conf : True -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] subcommand : push -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] quiet : False -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] cd_conf : -[2020-03-13 22:46:13,300][ceph_deploy.cli][INFO ] cluster : ceph -[2020-03-13 22:46:13,301][ceph_deploy.cli][INFO ] client : ['riot01', 'pine01', 'pine02', 'ebin01', 'ebin02', 'lenny', 'pine03'] -[2020-03-13 22:46:13,301][ceph_deploy.cli][INFO ] func : -[2020-03-13 22:46:13,301][ceph_deploy.cli][INFO ] ceph_conf : None -[2020-03-13 22:46:13,301][ceph_deploy.cli][INFO ] default_release : False -[2020-03-13 22:46:13,301][ceph_deploy.config][DEBUG ] Pushing config to riot01 -[2020-03-13 22:46:14,743][riot01][DEBUG ] connected to host: root@riot01 -[2020-03-13 22:46:14,744][riot01][DEBUG ] detect platform information from remote host -[2020-03-13 22:46:14,916][riot01][DEBUG ] detect machine type -[2020-03-13 22:46:14,951][riot01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:46:14,970][ceph_deploy.config][DEBUG ] Pushing config to pine01 -[2020-03-13 22:46:24,141][pine01][DEBUG ] connected to host: root@pine01 -[2020-03-13 22:46:24,143][pine01][DEBUG ] detect platform information from remote host -[2020-03-13 22:46:24,817][pine01][DEBUG ] detect machine type -[2020-03-13 22:46:24,892][pine01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:46:25,007][ceph_deploy.config][DEBUG ] Pushing config to pine02 -[2020-03-13 22:46:26,547][pine02][DEBUG ] connected to host: root@pine02 -[2020-03-13 22:46:26,549][pine02][DEBUG ] detect platform information from remote host -[2020-03-13 22:46:26,673][pine02][DEBUG ] detect machine type -[2020-03-13 22:46:26,693][pine02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:46:26,715][ceph_deploy.config][DEBUG ] Pushing config to ebin01 -[2020-03-13 22:46:29,415][ebin01][DEBUG ] connected to host: root@ebin01 -[2020-03-13 22:46:29,417][ebin01][DEBUG ] detect platform information from remote host -[2020-03-13 22:46:29,656][ebin01][DEBUG ] detect machine type -[2020-03-13 22:46:29,682][ebin01][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:46:29,701][ceph_deploy.config][DEBUG ] Pushing config to ebin02 -[2020-03-13 22:46:32,829][ebin02][DEBUG ] connected to host: root@ebin02 -[2020-03-13 22:46:32,831][ebin02][DEBUG ] detect platform information from remote host -[2020-03-13 22:46:33,104][ebin02][DEBUG ] detect machine type -[2020-03-13 22:46:33,136][ebin02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:46:33,162][ceph_deploy.config][DEBUG ] Pushing config to lenny -[2020-03-13 22:46:33,531][lenny][DEBUG ] connected to host: root@lenny -[2020-03-13 22:46:33,532][lenny][DEBUG ] detect platform information from remote host -[2020-03-13 22:46:33,548][lenny][DEBUG ] detect machine type -[2020-03-13 22:46:33,551][lenny][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf -[2020-03-13 22:46:33,553][ceph_deploy.config][DEBUG ] Pushing config to pine03 -[2020-03-13 22:46:44,406][pine03][DEBUG ] connected to host: root@pine03 -[2020-03-13 22:46:44,408][pine03][DEBUG ] detect platform information from remote host -[2020-03-13 22:46:44,882][pine03][DEBUG ] detect machine type -[2020-03-13 22:46:44,957][pine03][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf diff --git a/ceph.bootstrap-mds.keyring b/ceph.bootstrap-mds.keyring deleted file mode 100644 index 10c0b2e..0000000 --- a/ceph.bootstrap-mds.keyring +++ /dev/null @@ -1,3 +0,0 @@ -[client.bootstrap-mds] - key = AQAEtV5epuXwLhAALp4VAsSfVhs5XujznXBmUg== - caps mon = "allow profile bootstrap-mds" diff --git a/ceph.bootstrap-mgr.keyring b/ceph.bootstrap-mgr.keyring deleted file mode 100644 index f006a28..0000000 --- a/ceph.bootstrap-mgr.keyring +++ /dev/null @@ -1,3 +0,0 @@ -[client.bootstrap-mgr] - key = AQAEtV5eSV3xLhAAiwSninS8+gXUiK6AqcJ0jw== - caps mon = "allow profile bootstrap-mgr" diff --git a/ceph.bootstrap-osd.keyring b/ceph.bootstrap-osd.keyring deleted file mode 100644 index cf0fa85..0000000 --- a/ceph.bootstrap-osd.keyring +++ /dev/null @@ -1,3 +0,0 @@ -[client.bootstrap-osd] - key = AQAEtV5eaM7xLhAAOghmZDAD2g6s1knUyNNEjA== - caps mon = "allow profile bootstrap-osd" diff --git a/ceph.bootstrap-rgw.keyring b/ceph.bootstrap-rgw.keyring deleted file mode 100644 index 887eeba..0000000 --- a/ceph.bootstrap-rgw.keyring +++ /dev/null @@ -1,3 +0,0 @@ -[client.bootstrap-rgw] - key = AQAEtV5erUHzLhAAhhM/4/kbeOMzogETWZcs+A== - caps mon = "allow profile bootstrap-rgw" diff --git a/ceph.client.admin.keyring b/ceph.client.admin.keyring deleted file mode 100644 index 86910ad..0000000 --- a/ceph.client.admin.keyring +++ /dev/null @@ -1,6 +0,0 @@ -[client.admin] - key = AQAEtV5eTjjwLhAAjhBblIIarKdfZ01Y5wTkbg== - caps mds = "allow *" - caps mgr = "allow *" - caps mon = "allow *" - caps osd = "allow *" diff --git a/ceph.mon.keyring b/ceph.mon.keyring deleted file mode 100644 index 9b91ac1..0000000 --- a/ceph.mon.keyring +++ /dev/null @@ -1,3 +0,0 @@ -[mon.] -key = AQD7sF5eAAAAABAAJjvyccVMIhCSMPcPIihGCg== -caps mon = allow *