David, thanks.
I've switched the brnach to Luminous and the doc is the same (thankfully).

No worries, i'll wait till someone that hopefully did it already might give
me a hint.
thanks!

On Wed, Oct 11, 2017 at 11:00 AM, David Turner <drakonst...@gmail.com>
wrote:

> Careful when you're looking at documentation.  You're looking at the
> master branch which might have unreleased features or changes that your
> release doesn't have.  You'll want to change master in the url to luminous
> to make sure that you're looking at the documentation for your version of
> Ceph.
>
> I haven't personally used bluestore yet so I can't say what the proper
> commands are there without just looking online for the answer.  I do know
> that there is no reason to have your DB and WAL devices on separate
> partitions if they're on the same device.  What's been mentioned on the ML
> is that you want to create a partition for the DB and the WAL will use it.
> A partition for the WAL is only if it is planned to be on a different
> device than the DB.
>
> On Tue, Oct 10, 2017 at 5:59 PM Alejandro Comisario <alejan...@nubeliu.com>
> wrote:
>
>> Hi, i see some notes there that did'nt existed on jewel :
>>
>> http://docs.ceph.com/docs/master/rados/operations/add-
>> or-rm-osds/#replacing-an-osd
>>
>> In my case what im using right now on that OSD is this :
>>
>> root@ndc-cl-osd4:~# ls -lsah /var/lib/ceph/osd/ceph-104
>> total 64K
>>    0 drwxr-xr-x  2 ceph ceph  310 Sep 21 10:56 .
>> 4.0K drwxr-xr-x 25 ceph ceph 4.0K Sep 21 10:56 ..
>>    0 lrwxrwxrwx  1 ceph ceph   58 Sep 21 10:30 block ->
>> /dev/disk/by-partuuid/0ffa3ed7-169f-485c-9170-648ce656e9b1
>>    0 lrwxrwxrwx  1 ceph ceph   58 Sep 21 10:30 block.db ->
>> /dev/disk/by-partuuid/5873e2cb-3c26-4a7d-8ff1-1bc3e2d62e5a
>>    0 lrwxrwxrwx  1 ceph ceph   58 Sep 21 10:30 block.wal ->
>> /dev/disk/by-partuuid/aed9e5e4-c798-46b5-8243-e462e74f6485
>>
>> block.db and block.wal are on two different NVME partitions, witch are 
>> nvme1n1p17
>> and nvme1n1p18 so assuming after hot swaping the device, the drive
>> letter is "sdx" according to the link above what would be the right command
>> to re-use the two NVME partitions for block db and wal ?
>>
>> I presume that everything else is the same.
>> best.
>>
>>
>> On Sat, Sep 30, 2017 at 9:00 PM, David Turner <drakonst...@gmail.com>
>> wrote:
>>
>>> I'm pretty sure that the process is the same as with filestore. The
>>> cluster doesn't really know if an osd is filestore or bluestore... It's
>>> just an osd running a daemon.
>>>
>>> If there are any differences, they would be in the release notes for
>>> Luminous as changes from Jewel.
>>>
>>> On Sat, Sep 30, 2017, 6:28 PM Alejandro Comisario <alejan...@nubeliu.com>
>>> wrote:
>>>
>>>> Hi all.
>>>> Independetly that i've deployerd a ceph Luminous cluster with Bluestore
>>>> using ceph-ansible (https://github.com/ceph/ceph-ansible) what is the
>>>> right way to replace a disk when using Bluestore ?
>>>>
>>>> I will try to forget everything i know on how to recover things with
>>>> filestore and start fresh.
>>>>
>>>> Any how-to's ? experiences ? i dont seem to find an official way of
>>>> doing it.
>>>> best.
>>>>
>>>> --
>>>> *Alejandro Comisario*
>>>> *CTO | NUBELIU*
>>>> E-mail: alejandro@nubeliu.comCell: +54 9 11 3770 1857
>>>> <+54%209%2011%203770-1857>
>>>> _
>>>> www.nubeliu.com
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>
>>
>>
>> --
>> *Alejandro Comisario*
>> *CTO | NUBELIU*
>> E-mail: alejandro@nubeliu.comCell: +54911 3770 1857
>> _
>>
>


-- 
*Alejandro Comisario*
*CTO | NUBELIU*
E-mail: alejandro@nubeliu.comCell: +54911 3770 1857
_
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to