On Fri, Jun 8, 2018 at 3:59 PM, Rares Vernica <rvern...@gmail.com> wrote:
> Thanks, I will try that. > > Just to verify I don't need to create a file system or any partition table > on the volume, right? Ceph seems to be trying to create the file system. > Right, no need to do anything here for filesystems. > > On Fri, Jun 8, 2018 at 12:56 PM, Alfredo Deza <ad...@redhat.com> wrote: > >> >> >> On Fri, Jun 8, 2018 at 3:17 PM, Rares Vernica <rvern...@gmail.com> wrote: >> >>> Yes, it exists: >>> >>> # ls -ld /var/lib/ceph/osd/ceph-0 >>> drwxr-xr-x. 2 ceph ceph 6 Jun 7 15:06 /var/lib/ceph/osd/ceph-0 >>> # ls -ld /var/lib/ceph/osd >>> drwxr-x---. 4 ceph ceph 34 Jun 7 15:59 /var/lib/ceph/osd >>> >>> After I ran the ceph-volume command, I see the directory is mounted: >>> >>> # mount >>> ... >>> tmpfs on /var/lib/ceph/osd/ceph-0 type tmpfs (rw,relatime,seclabel) >>> # ls -ld /var/lib/ceph/osd/ceph-0 >>> drwxrwxrwt. 2 ceph ceph 160 Jun 8 12:15 /var/lib/ceph/osd/ceph-0 >>> >>> >>> >> I don't know what to say. The only other thing that jumps at me is that >> this is a 1GB device. Have you tried with other sizes? You could also try >> with a whole raw device and let >> ceph-volume create the vg/lv for you: ceph-volume lvm create --data >> /dev/<mydevice> >> >> >> >> >>> Thanks, >>> Rares >>> >>> >>> >>> On Fri, Jun 8, 2018 at 12:09 PM, Alfredo Deza <ad...@redhat.com> wrote: >>> >>>> >>>> >>>> On Fri, Jun 8, 2018 at 2:47 PM, Rares Vernica <rvern...@gmail.com> >>>> wrote: >>>> >>>>> Hi, >>>>> >>>>> I'm following the Manual Deployment guide at >>>>> http://docs.ceph.com/docs/master/install/manual-deployment/ I'm not >>>>> able to move past the ceph-volume lvm create part. Here is what I do: >>>>> >>>>> # lvcreate -L 1G -n ceph cah_foo >>>>> Logical volume "ceph" created. >>>>> >>>>> # ceph-volume lvm create --data cah_foo/ceph >>>>> Running command: /bin/ceph-authtool --gen-print-key >>>>> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd >>>>> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new >>>>> b8016385-e46c-4e93-a334-be4fc92bea85 >>>>> Running command: /bin/ceph-authtool --gen-print-key >>>>> Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 >>>>> Running command: chown -R ceph:ceph /dev/dm-2 >>>>> Running command: ln -s /dev/cah_foo/ceph /var/lib/ceph/osd/ceph-0/block >>>>> Running command: ceph --cluster ceph --name client.bootstrap-osd >>>>> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o >>>>> /var/lib/ceph/osd/ceph-0/activate.monmap >>>>> stderr: got monmap epoch 2 >>>>> Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring >>>>> --create-keyring --name osd.0 --add-key AQCxuhlbAVylMRAAXsKQpKbau3T1rI >>>>> 66z651ng== >>>>> stdout: creating /var/lib/ceph/osd/ceph-0/keyring >>>>> added entity osd.0 auth auth(auid = 18446744073709551615 >>>>> key=AQCxuhlbAVylMRAAXsKQpKbau3T1rI66z651ng== with 0 caps) >>>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring >>>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ >>>>> Running command: /bin/ceph-osd --cluster ceph --osd-objectstore >>>>> bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap >>>>> --keyfile - --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid >>>>> b8016385-e46c-4e93-a334-be4fc92bea85 --setuser ceph --setgroup ceph >>>>> stderr: 2018-06-07 16:07:32.804440 7f237709dd80 -1 >>>>> bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid >>>>> stderr: 2018-06-07 16:07:33.822761 7f237709dd80 -1 OSD::mkfs: >>>>> ObjectStore::mkfs failed with error (2) No such file or directory >>>>> stderr: 2018-06-07 16:07:33.822934 7f237709dd80 -1 *** ERROR: error >>>>> creating empty object store in /var/lib/ceph/osd/ceph-0/: (2) No such file >>>>> or directory* >>>>> --> ceph-volume lvm prepare successful for: cah_foo/ceph >>>>> Running command: ceph-bluestore-tool --cluster=ceph prime-osd-dir >>>>> --dev /dev/cah_foo/ceph --path /var/lib/ceph/osd/ceph-0 >>>>> stderr: failed to read label for /dev/cah_foo/ceph: (2) No such file >>>>> or directory >>>>> --> Was unable to complete a new OSD, will rollback changes >>>>> --> OSD will be fully purged from the cluster, because the ID was >>>>> generated >>>>> Running command: ceph osd purge osd.0 --yes-i-really-mean-it >>>>> stderr: purged osd.0 >>>>> --> RuntimeError: command returned non-zero exit status: 1 >>>>> >>>>> # ceph --version >>>>> ceph version 12.2.5 (cad919881333ac92274171586c827e01f554a70a) >>>>> luminous (stable) >>>>> >>>> >>>> This looks really odd. Do you have a /var/lib/ceph/osd/ceph-0 >>>> directory? if yes, what are the permissions inside of it? The output shows >>>> that it is mounting this correctly: >>>> >>>> Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 >>>> >>>> So that should exist >>>> >>>>> >>>>> >>>>> I wonder what am I missing and what else I can try. >>>>> >>>>> Thanks! >>>>> Rares >>>>> >>>>> >>>>> _______________________________________________ >>>>> ceph-users mailing list >>>>> ceph-users@lists.ceph.com >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >>>>> >>>> >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>> >> > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com