Thanks, I will try that.

Just to verify I don't need to create a file system or any partition table
on the volume, right? Ceph seems to be trying to create the file system.

On Fri, Jun 8, 2018 at 12:56 PM, Alfredo Deza <ad...@redhat.com> wrote:

>
>
> On Fri, Jun 8, 2018 at 3:17 PM, Rares Vernica <rvern...@gmail.com> wrote:
>
>> Yes, it exists:
>>
>> # ls -ld /var/lib/ceph/osd/ceph-0
>> drwxr-xr-x. 2 ceph ceph 6 Jun  7 15:06 /var/lib/ceph/osd/ceph-0
>> # ls -ld /var/lib/ceph/osd
>> drwxr-x---. 4 ceph ceph 34 Jun  7 15:59 /var/lib/ceph/osd
>>
>> After I ran the ceph-volume command, I see the directory is mounted:
>>
>> # mount
>> ...
>> tmpfs on /var/lib/ceph/osd/ceph-0 type tmpfs (rw,relatime,seclabel)
>> # ls -ld /var/lib/ceph/osd/ceph-0
>> drwxrwxrwt. 2 ceph ceph 160 Jun  8 12:15 /var/lib/ceph/osd/ceph-0
>>
>>
>>
> I don't know what to say. The only other thing that jumps at me is that
> this is a 1GB device. Have you tried with other sizes? You could also try
> with a whole raw device and let
> ceph-volume create the vg/lv for you:  ceph-volume lvm create --data
> /dev/<mydevice>
>
>
>
>
>> Thanks,
>> Rares
>>
>>
>>
>> On Fri, Jun 8, 2018 at 12:09 PM, Alfredo Deza <ad...@redhat.com> wrote:
>>
>>>
>>>
>>> On Fri, Jun 8, 2018 at 2:47 PM, Rares Vernica <rvern...@gmail.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> I'm following the Manual Deployment guide at
>>>> http://docs.ceph.com/docs/master/install/manual-deployment/ I'm not
>>>> able to move past the ceph-volume lvm create part. Here is what I do:
>>>>
>>>> # lvcreate -L 1G -n ceph cah_foo
>>>>   Logical volume "ceph" created.
>>>>
>>>> # ceph-volume lvm create --data cah_foo/ceph
>>>> Running command: /bin/ceph-authtool --gen-print-key
>>>> Running command: /bin/ceph --cluster ceph --name client.bootstrap-osd
>>>> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
>>>> b8016385-e46c-4e93-a334-be4fc92bea85
>>>> Running command: /bin/ceph-authtool --gen-print-key
>>>> Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0
>>>> Running command: chown -R ceph:ceph /dev/dm-2
>>>> Running command: ln -s /dev/cah_foo/ceph /var/lib/ceph/osd/ceph-0/block
>>>> Running command: ceph --cluster ceph --name client.bootstrap-osd
>>>> --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o
>>>> /var/lib/ceph/osd/ceph-0/activate.monmap
>>>>  stderr: got monmap epoch 2
>>>> Running command: ceph-authtool /var/lib/ceph/osd/ceph-0/keyring
>>>> --create-keyring --name osd.0 --add-key AQCxuhlbAVylMRAAXsKQpKbau3T1rI
>>>> 66z651ng==
>>>>  stdout: creating /var/lib/ceph/osd/ceph-0/keyring
>>>> added entity osd.0 auth auth(auid = 18446744073709551615
>>>> key=AQCxuhlbAVylMRAAXsKQpKbau3T1rI66z651ng== with 0 caps)
>>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring
>>>> Running command: chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/
>>>> Running command: /bin/ceph-osd --cluster ceph --osd-objectstore
>>>> bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap
>>>> --keyfile - --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid
>>>> b8016385-e46c-4e93-a334-be4fc92bea85 --setuser ceph --setgroup ceph
>>>>  stderr: 2018-06-07 16:07:32.804440 7f237709dd80 -1
>>>> bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid
>>>>  stderr: 2018-06-07 16:07:33.822761 7f237709dd80 -1 OSD::mkfs:
>>>> ObjectStore::mkfs failed with error (2) No such file or directory
>>>>  stderr: 2018-06-07 16:07:33.822934 7f237709dd80 -1  *** ERROR: error
>>>> creating empty object store in /var/lib/ceph/osd/ceph-0/: (2) No such file
>>>> or directory*
>>>> --> ceph-volume lvm prepare successful for: cah_foo/ceph
>>>> Running command: ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev
>>>> /dev/cah_foo/ceph --path /var/lib/ceph/osd/ceph-0
>>>>  stderr: failed to read label for /dev/cah_foo/ceph: (2) No such file
>>>> or directory
>>>> --> Was unable to complete a new OSD, will rollback changes
>>>> --> OSD will be fully purged from the cluster, because the ID was
>>>> generated
>>>> Running command: ceph osd purge osd.0 --yes-i-really-mean-it
>>>>  stderr: purged osd.0
>>>> -->  RuntimeError: command returned non-zero exit status: 1
>>>>
>>>> # ceph --version
>>>> ceph version 12.2.5 (cad919881333ac92274171586c827e01f554a70a)
>>>> luminous (stable)
>>>>
>>>
>>> This looks really odd. Do you have a /var/lib/ceph/osd/ceph-0 directory?
>>> if yes, what are the permissions inside of it? The output shows that it is
>>> mounting this correctly:
>>>
>>> Running command: mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0
>>>
>>> So that should exist
>>>
>>>>
>>>>
>>>> I wonder what am I missing and what else I can try.
>>>>
>>>> Thanks!
>>>> Rares
>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>>
>>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to