[ceph-users] ceph osd down

2016-11-20 Thread 马忠明
Hi guys, So our cluster always got osd down due to medium error.Our current action plan is to replace the defective disk drive.But I was wondering whether it's too sensitive for ceph to take it down.Or whether our action plan was too simple and crude.Any advice for this issue will be appreciated

Re: [ceph-users] ceph OSD down+out =>health ok => remove =>PGsbackfilling... ?

2016-04-26 Thread Burkhard Linke
Hi, On 04/26/2016 12:32 PM, SCHAER Frederic wrote: Hi, One simple/quick question. In my ceph cluster, I had a disk wich was in predicted failure. It was so much in predicted failure that the ceph OSD daemon crashed. After the OSD crashed, ceph moved data correctly (or at least that’s what

[ceph-users] ceph OSD down+out =>health ok => remove => PGs backfilling... ?

2016-04-26 Thread SCHAER Frederic
Hi, One simple/quick question. In my ceph cluster, I had a disk wich was in predicted failure. It was so much in predicted failure that the ceph OSD daemon crashed. After the OSD crashed, ceph moved data correctly (or at least that's what I thought), and a ceph -s was giving a "HEALTH_OK". Perf

Re: [ceph-users] ceph osd down and out

2014-06-05 Thread Cao, Buddy
7f2ce92cb700 time > 2014-06-05 > 10:27:54.703693 Wei Cao (Buddy) -Original Message- From: Cao, Buddy Sent: Thursday, June 5, 2014 11:19 PM To: 'Sage Weil' Cc: ceph-users@lists.ceph.com Subject: RE: [ceph-users] ceph osd down and out Sage, Yes, I already set the max o

Re: [ceph-users] ceph osd down and out

2014-06-05 Thread Cao, Buddy
Cao (Buddy) -Original Message- From: Sage Weil [mailto:s...@inktank.com] Sent: Thursday, June 5, 2014 11:11 PM To: Cao, Buddy Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] ceph osd down and out This usually happens on larger clusters when you hit the max fd limit. Add

Re: [ceph-users] ceph osd down and out

2014-06-05 Thread Sage Weil
This usually happens on larger clusters when you hit the max fd limit. Add max open files = 131072 in the [global] section of ceph.conf to fix it (default is 16384). sage On Thu, 5 Jun 2014, Cao, Buddy wrote: > > Hi,  several osds were down/out with similar logs as below, could you

[ceph-users] ceph osd down and out

2014-06-05 Thread Cao, Buddy
Hi, several osds were down/out with similar logs as below, could you help? -38> 2014-06-05 10:27:54.700832 7f2ceead6700 1 -- 192.168.40.11:6800/19542 <== osd.11 192.168.40.11:6822/20298 2 pg_notify(0.aa4(2) epoch 7) v5 812+0+0 (3873498789 0 0) 0x57a0540 con 0x49d14a0 -37> 2014-0