---------- Forwarded message ---------- From: "hp cre" <hpc...@gmail.com> Date: 11 Apr 2016 15:50 Subject: Re: [ceph-users] Ubuntu xenial and ceph jewel systemd To: "James Page" <james.p...@ubuntu.com> Cc:
Here is exactly what has been done (just started from scratch today): 1- install default xenial beta 2 2- run apt-get update && apt-get dist-upgrade (this step was not done on first trial) after update, got warning as follows: "W: plymouth: The plugin label.so is missing, the selected theme might not work as expected. W: plymouth: You might want to install the plymouth-themes and plymouth-label package to fix this. W: mdadm: /etc/mdadm/mdadm.conf defines no arrays. " so i ran apt-get -y install plymouth-themes 3- wget http://download.ceph.com/debian-jewel/pool/main/c/ceph-deploy/ceph-deploy_1.5.31_all.deb 4- dpkg -i ceph-deploy_1.5.31_all.deb got errors of unmet dependencies, so i ran apt-get -f install. this installed all missing packages. 5- followed ceph docs preflight checklist (sudo file, ssh config file, ssh-copy-id, install ntp) Followed the storage cluster quick start guide 6- ceph-deploy new xen1 (first node) --> all ok 7- edit ceph.conf --> osd pool default size = 2 8- ceph-deploy install --release=jewel xen1 --> all ok (this time it installed jewel 10.1.1, yesterday it was 10.1.0) 9- ceph-deploy mon create-initial --> same error: wes@xen1:~/cl$ ceph-deploy mon create-initial [ceph_deploy.conf][DEBUG ] found configuration file at: /home/wes/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (1.5.31): /usr/bin/ceph-deploy mon create-initial [ceph_deploy.cli][INFO ] ceph-deploy options: [ceph_deploy.cli][INFO ] username : None [ceph_deploy.cli][INFO ] verbose : False [ceph_deploy.cli][INFO ] overwrite_conf : False [ceph_deploy.cli][INFO ] subcommand : create-initial [ceph_deploy.cli][INFO ] quiet : False [ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7ffb88bfed88> [ceph_deploy.cli][INFO ] cluster : ceph [ceph_deploy.cli][INFO ] func : <function mon at 0x7ffb88bdcf50> [ceph_deploy.cli][INFO ] ceph_conf : None [ceph_deploy.cli][INFO ] keyrings : None [ceph_deploy.cli][INFO ] default_release : False [ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts xen1 [ceph_deploy.mon][DEBUG ] detecting platform for host xen1 ... [xen1][DEBUG ] connection detected need for sudo [xen1][DEBUG ] connected to host: xen1 [xen1][DEBUG ] detect platform information from remote host [xen1][DEBUG ] detect machine type [xen1][DEBUG ] find the location of an executable [ceph_deploy.mon][INFO ] distro info: Ubuntu 16.04 xenial [xen1][DEBUG ] determining if provided host has same hostname in remote [xen1][DEBUG ] get remote short hostname [xen1][DEBUG ] deploying mon to xen1 [xen1][DEBUG ] get remote short hostname [xen1][DEBUG ] remote hostname: xen1 [xen1][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf [xen1][DEBUG ] create the mon path if it does not exist [xen1][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-xen1/done [xen1][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-xen1/done [xen1][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-xen1.mon.keyring [xen1][DEBUG ] create the monitor keyring file [xen1][INFO ] Running command: sudo ceph-mon --cluster ceph --mkfs -i xen1 --keyring /var/lib/ceph/tmp/ceph-xen1.mon.keyring --setuser 64045 --setgroup 64045 [xen1][DEBUG ] ceph-mon: mon.noname-a 192.168.56.10:6789/0 is local, renaming to mon.xen1 [xen1][DEBUG ] ceph-mon: set fsid to d56c2ad9-dc66-4b6a-b269-e32eecc05571 [xen1][DEBUG ] ceph-mon: created monfs at /var/lib/ceph/mon/ceph-xen1 for mon.xen1 [xen1][INFO ] unlinking keyring file /var/lib/ceph/tmp/ceph-xen1.mon.keyring [xen1][DEBUG ] create a done file to avoid re-doing the mon deployment [xen1][DEBUG ] create the init path if it does not exist [xen1][INFO ] Running command: sudo initctl emit ceph-mon cluster=ceph id=xen1 [xen1][ERROR ] Traceback (most recent call last): [xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/process.py", line 119, in run [xen1][ERROR ] reporting(conn, result, timeout) [xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/log.py", line 13, in reporting [xen1][ERROR ] received = result.receive(timeout) [xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/lib/vendor/execnet/gateway_base.py", line 704, in receive [xen1][ERROR ] raise self._getremoteerror() or EOFError() [xen1][ERROR ] RemoteError: Traceback (most recent call last): [xen1][ERROR ] File "/usr/lib/python2.7/dist-packages/ceph_deploy/lib/vendor/remoto/lib/vendor/execnet/gateway_base.py", line 1036, in executetask [xen1][ERROR ] function(channel, **kwargs) [xen1][ERROR ] File "<remote exec>", line 12, in _remote_run [xen1][ERROR ] File "/usr/lib/python2.7/subprocess.py", line 711, in __init__ [xen1][ERROR ] errread, errwrite) [xen1][ERROR ] File "/usr/lib/python2.7/subprocess.py", line 1340, in _execute_child [xen1][ERROR ] raise child_exception [xen1][ERROR ] OSError: [Errno 2] No such file or directory [xen1][ERROR ] [xen1][ERROR ] [ceph_deploy.mon][ERROR ] Failed to execute command: initctl emit ceph-mon cluster=ceph id=xen1 [ceph_deploy][ERROR ] GenericError: Failed to create 1 monitors On 11 April 2016 at 14:05, hp cre <hpc...@gmail.com> wrote: > In the process of reproducing it now. I'll attach a full command log > On 11 Apr 2016 11:42, "James Page" <james.p...@ubuntu.com> wrote: > > It would be handy to get visibility of your deployment log data; I'm not > currently able to reproduce your issue deploying ceph using ceph-deploy on > a small three node install running xenial - its correctly detecting systemd > and using systemctl instead of initctl. > > On Mon, 11 Apr 2016 at 10:18 James Page <james.p...@ubuntu.com> wrote: > >> On Mon, 11 Apr 2016 at 10:02 hp cre <hpc...@gmail.com> wrote: >> >>> Hello James, >>> >>> It's a default install of xenial server beta 2 release. Created a user >>> then followed the ceph installation quick start exactly as it is. >>> >>> Ceph-deploy version 1.5.31 was used as follows >>> >>> 1- ceph-deploy new node1 >>> 2- ceph-deploy install --release jewel node1 >>> 3- ceph-deploy mon create-initial >>> >>> Step 3 gave error in Python scripts. Meaning it could not find initctl >>> command. Searched for this command and found out our belongs to upstart. >>> >> I suspect that ceph-deploy is not playing nicely with systemd based >> Ubuntu releases - I'll take a look now.. >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com