Thanks, I will try that. Just to verify I don't need to create a file system or any partition table on the volume, right? Ceph seems to be trying to create the file system.
On Fri, Jun 8, 2018 at 12:56 PM, Alfredo Deza <ad...@redhat.com> wrote: > > > On Fri, Jun 8, 2018 at 3:17 PM, Rares Vernica <rvern...@gmail.com> wrote: > >> Yes, it exists: >> >> # ls -ld /var/lib/ceph/osd/ceph-0 >> drwxr-xr-x. 2 ceph ceph 6 Jun 7 15:06 /var/lib/ceph/osd/ceph-0 >> # ls -ld /var/lib/ceph/osd >> drwxr-x---. 4 ceph ceph 34 Jun 7 15:59 /var/lib/ceph/osd >> >> After I ran the ceph-volume command, I see the directory is mounted: >> >> # mount >> ... >> tmpfs on /var/lib/ceph/osd/ceph-0 type tmpfs (rw,relatime,seclabel) >> # ls -ld /var/lib/ceph/osd/ceph-0 >> drwxrwxrwt. 2 ceph ceph 160 Jun 8 12:15 /var/lib/ceph/osd/ceph-0 >> >> >> > I don't know what to say. The only other thing that jumps at me is that > this is a 1GB device. Have you tried with other sizes? You could also try > with a whole raw device and let > ceph-volume create the vg/lv for you: ceph-volume lvm create --data > /dev/<mydevice> > > > > >> Thanks, >> Rares >> >> >> >> On Fri, Jun 8, 2018 at 12:09 PM, Alfredo Deza <ad...@redhat.com> wrote: >> >>> >>> >>> On Fri, Jun 8, 2018 at 2:47 PM, Rares Vernica <rvern...@gmail.com> >>> wrote: >>> >>>> Hi, >>>> >>>> I'm following the Manual Deployment guide at >>>> http://docs.ceph.com/docs/master/install/manual-deployment/ I'm not >>>> able to move past the ceph-volume lvm create part. Here is what I do: >>>> >>>> # lvcreate -L 1G -n ceph cah_foo >>>> Logical volume "ceph" created. >>>> >>>> # ceph-volume lvm create --data cah_foo/ceph >>>> Running command: /bin/ceph-authtool --gen-print-key >>>> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd >>>> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new >>>> b8016385-e46c-4e93-a334-be4fc92bea85 >>>> Running command: /bin/ceph-authtool --gen-print-key >>>> Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 >>>> Running command: chown -R ceph:ceph /dev/dm-2 >>>> Running command: ln -s /dev/cah_foo/ceph /var/lib/ceph/osd/ceph-0/block >>>> Running command: ceph --cluster ceph --name client.bootstrap-osd >>>> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o >>>> /var/lib/ceph/osd/ceph-0/activate.monmap >>>> stderr: got monmap epoch 2 >>>> Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring >>>> --create-keyring --name osd.0 --add-key AQCxuhlbAVylMRAAXsKQpKbau3T1rI >>>> 66z651ng== >>>> stdout: creating /var/lib/ceph/osd/ceph-0/keyring >>>> added entity osd.0 auth auth(auid = 18446744073709551615 >>>> key=AQCxuhlbAVylMRAAXsKQpKbau3T1rI66z651ng== with 0 caps) >>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring >>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/ >>>> Running command: /bin/ceph-osd --cluster ceph --osd-objectstore >>>> bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap >>>> --keyfile - --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid >>>> b8016385-e46c-4e93-a334-be4fc92bea85 --setuser ceph --setgroup ceph >>>> stderr: 2018-06-07 16:07:32.804440 7f237709dd80 -1 >>>> bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid >>>> stderr: 2018-06-07 16:07:33.822761 7f237709dd80 -1 OSD::mkfs: >>>> ObjectStore::mkfs failed with error (2) No such file or directory >>>> stderr: 2018-06-07 16:07:33.822934 7f237709dd80 -1 *** ERROR: error >>>> creating empty object store in /var/lib/ceph/osd/ceph-0/: (2) No such file >>>> or directory* >>>> --> ceph-volume lvm prepare successful for: cah_foo/ceph >>>> Running command: ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev >>>> /dev/cah_foo/ceph --path /var/lib/ceph/osd/ceph-0 >>>> stderr: failed to read label for /dev/cah_foo/ceph: (2) No such file >>>> or directory >>>> --> Was unable to complete a new OSD, will rollback changes >>>> --> OSD will be fully purged from the cluster, because the ID was >>>> generated >>>> Running command: ceph osd purge osd.0 --yes-i-really-mean-it >>>> stderr: purged osd.0 >>>> --> RuntimeError: command returned non-zero exit status: 1 >>>> >>>> # ceph --version >>>> ceph version 12.2.5 (cad919881333ac92274171586c827e01f554a70a) >>>> luminous (stable) >>>> >>> >>> This looks really odd. Do you have a /var/lib/ceph/osd/ceph-0 directory? >>> if yes, what are the permissions inside of it? The output shows that it is >>> mounting this correctly: >>> >>> Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0 >>> >>> So that should exist >>> >>>> >>>> >>>> I wonder what am I missing and what else I can try. >>>> >>>> Thanks! >>>> Rares >>>> >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> >>> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com