Hi guys,
So our cluster always got osd down due to medium error.Our current action plan
is to replace the defective disk drive.But I was wondering whether it's too
sensitive for ceph to take it down.Or whether our action plan was too simple
and crude.Any advice for this issue will be appreciated
Hi,
On 04/26/2016 12:32 PM, SCHAER Frederic wrote:
Hi,
One simple/quick question.
In my ceph cluster, I had a disk wich was in predicted failure. It was
so much in predicted failure that the ceph OSD daemon crashed.
After the OSD crashed, ceph moved data correctly (or at least that’s
what
Hi,
One simple/quick question.
In my ceph cluster, I had a disk wich was in predicted failure. It was so much
in predicted failure that the ceph OSD daemon crashed.
After the OSD crashed, ceph moved data correctly (or at least that's what I
thought), and a ceph -s was giving a "HEALTH_OK".
Perf
7f2ce92cb700 time
> 2014-06-05
> 10:27:54.703693
Wei Cao (Buddy)
-Original Message-
From: Cao, Buddy
Sent: Thursday, June 5, 2014 11:19 PM
To: 'Sage Weil'
Cc: ceph-users@lists.ceph.com
Subject: RE: [ceph-users] ceph osd down and out
Sage,
Yes, I already set the max o
Cao (Buddy)
-Original Message-
From: Sage Weil [mailto:s...@inktank.com]
Sent: Thursday, June 5, 2014 11:11 PM
To: Cao, Buddy
Cc: ceph-users@lists.ceph.com
Subject: Re: [ceph-users] ceph osd down and out
This usually happens on larger clusters when you hit the max fd limit.
Add
This usually happens on larger clusters when you hit the max fd limit.
Add
max open files = 131072
in the [global] section of ceph.conf to fix it (default is 16384).
sage
On Thu, 5 Jun 2014, Cao, Buddy wrote:
>
> Hi, several osds were down/out with similar logs as below, could you
Hi, several osds were down/out with similar logs as below, could you help?
-38> 2014-06-05 10:27:54.700832 7f2ceead6700 1 -- 192.168.40.11:6800/19542
<== osd.11 192.168.40.11:6822/20298 2 pg_notify(0.aa4(2) epoch 7) v5
812+0+0 (3873498789 0 0) 0x57a0540 con 0x49d14a0
-37> 2014-0