I believe you are absolutelly right.
It was my fault not checking the dates before posting, my bad.

Thanks for you help.
best.

On Tue, Oct 17, 2017 at 8:14 PM, Jamie Fargen <jfar...@redhat.com> wrote:

> Alejandro-
>
> Those are kernel messages indicating that the an error was encountered
> when data was sent to the storage device and are not related directly to
> the operation of Ceph. The messages you sent also appear to have happened 4
> days ago on Friday and if they have subsided then it probably means nothing
> further has tried to read/write to the disk, but the messages will be
> present in dmesg until the kernel ring buffer is overwritten or the system
> is restarted.
>
> -Jamie
>
>
> On Tue, Oct 17, 2017 at 6:47 PM, Alejandro Comisario <
> alejan...@nubeliu.com> wrote:
>
>> Jamie, thanks for replying, info is as follow:
>>
>> 1)
>>
>> [Fri Oct 13 10:21:24 2017] sd 0:2:23:0: [sdx] tag#0 FAILED Result:
>> hostbyte=DID_OK driverbyte=DRIVER_SENSE
>> [Fri Oct 13 10:21:24 2017] sd 0:2:23:0: [sdx] tag#0 Sense Key : Medium
>> Error [current]
>> [Fri Oct 13 10:21:24 2017] sd 0:2:23:0: [sdx] tag#0 Add. Sense: No
>> additional sense information
>> [Fri Oct 13 10:21:24 2017] sd 0:2:23:0: [sdx] tag#0 CDB: Read(10) 28 00
>> 00 00 09 10 00 00 f0 00
>> [Fri Oct 13 10:21:24 2017] blk_update_request: I/O error, dev sdx, sector
>> 2320
>>
>> 2)
>>
>> ndc-cl-mon1:~# ceph status
>>   cluster:
>>     id:     48158350-ba8a-420b-9c09-68da57205924
>>     health: HEALTH_OK
>>
>>   services:
>>     mon: 3 daemons, quorum ndc-cl-mon1,ndc-cl-mon2,ndc-cl-mon3
>>     mgr: ndc-cl-mon1(active), standbys: ndc-cl-mon3, ndc-cl-mon2
>>     osd: 161 osds: 160 up, 160 in
>>
>>   data:
>>     pools:   4 pools, 12288 pgs
>>     objects: 663k objects, 2650 GB
>>     usage:   9695 GB used, 258 TB / 267 TB avail
>>     pgs:     12288 active+clean
>>
>>   io:
>>     client:   0 B/s rd, 1248 kB/s wr, 49 op/s rd, 106 op/s wr
>>
>> 3)
>>
>> https://pastebin.com/MeCKqvp1
>>
>>
>> On Tue, Oct 17, 2017 at 5:59 PM, Jamie Fargen <jfar...@redhat.com> wrote:
>>
>>> Alejandro-
>>> Please provide the folloing information:
>>> 1) Include an example of an actual message you are seeing in dmesg.
>>> 2) Provide the output of # ceph status
>>> 3) Provide the output of # ceph osd tree
>>>
>>> Regards,
>>> Jamie Fargen
>>>
>>>
>>>
>>> On Tue, Oct 17, 2017 at 4:34 PM, Alejandro Comisario <
>>> alejan...@nubeliu.com> wrote:
>>>
>>>> hi guys, any tip or help ?
>>>>
>>>> On Mon, Oct 16, 2017 at 1:50 PM, Alejandro Comisario <
>>>> alejan...@nubeliu.com> wrote:
>>>>
>>>>> Hi all, i have to hot-swap a failed osd on a Luminous Cluster with
>>>>> Blue store (the disk is SATA, WAL and DB are on NVME).
>>>>>
>>>>> I've issued a:
>>>>> * ceph osd crush reweight osd_id 0
>>>>> * systemctl stop (osd I'd daemon)
>>>>> * umount /var/lib/ceph/osd/osd_id
>>>>> * ceph osd destroy osd_id
>>>>>
>>>>> everything seems of, but if I left everything as is ( until I wait for
>>>>> the replaced disk ) I can see that dmesg errors on writing on the device
>>>>> are still appearing.
>>>>>
>>>>> The osd is of course down and out the crushmap.
>>>>> am I missing something ? like a step to execute or something else ?
>>>>>
>>>>> hoping to get help.
>>>>> best.
>>>>>
>>>>> ​alejandrito
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> *Alejandro Comisario*
>>>> *CTO | NUBELIU*
>>>> E-mail: alejandro@nubeliu.comCell: +54911 3770 1857
>>>> _
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>>
>>>
>>>
>>> --
>>> Jamie Fargen
>>> Consultant
>>> jfar...@redhat.com
>>> 813-817-4430 <(813)%20817-4430>
>>>
>>
>>
>>
>> --
>> *Alejandro Comisario*
>> *CTO | NUBELIU*
>> E-mail: alejandro@nubeliu.comCell: +54911 3770 1857
>> _
>>
>
>
>
> --
> Jamie Fargen
> Consultant
> jfar...@redhat.com
> 813-817-4430
>



-- 
*Alejandro Comisario*
*CTO | NUBELIU*
E-mail: alejandro@nubeliu.comCell: +54911 3770 1857
_
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to