[ceph-users] HEALTH_WARN after upgrade to cuttlefish

2013-05-08 Thread James Harper
I've just upgraded my ceph install to cuttlefish (was 0.60) from Debian. My mon's don't regularly die anymore, or at least haven't so far, but health is always HEALTH_WARN even though I can't see any indication of why: # ceph status health HEALTH_WARN monmap e1: 3 mons at {4=192.168.200.1

Re: [ceph-users] HEALTH_WARN after upgrade to cuttlefish

2013-05-08 Thread David Zafman
According to "osdmap e504: 4 osds: 2 up, 2 in" you have 2 of 4 osds that are down and out. That may be the issue. David Zafman Senior Developer http://www.inktank.com On May 8, 2013, at 12:05 AM, James Harper wrote: > I've just upgraded my ceph install to cuttlefish (was 0.60) from Debian. >

Re: [ceph-users] HEALTH_WARN after upgrade to cuttlefish

2013-05-08 Thread Joao Eduardo Luis
On 05/08/2013 08:44 AM, David Zafman wrote: According to "osdmap e504: 4 osds: 2 up, 2 in" you have 2 of 4 osds that are down and out. That may be the issue. Also, running 'ceph health detail' will give you specifics on what is causing the HEALTH_WARN. -Joao David Zafman Senior Devel

Re: [ceph-users] HEALTH_WARN after upgrade to cuttlefish

2013-05-08 Thread James Harper
> On 05/08/2013 08:44 AM, David Zafman wrote: > > > > According to "osdmap e504: 4 osds: 2 up, 2 in" you have 2 of 4 osds that are > down and out. That may be the issue. > > Also, running 'ceph health detail' will give you specifics on what is > causing the HEALTH_WARN. > # ceph health detail H

Re: [ceph-users] Dell R515 performance and specification question

2013-05-08 Thread Barry O'Rourke
Hi, I've been doing some numbers today and it looks like our choice is between 6 x R515's or 6 x R410's depending upon whether we want to allow for the possibility of adding more OSDs at a later date. Do you have any experience with the Dell H200 cards? You mentioned earlier that the Dell S

Re: [ceph-users] Dell R515 performance and specification question

2013-05-08 Thread Mark Nelson
On 05/08/2013 07:08 AM, Barry O'Rourke wrote: Hi, I've been doing some numbers today and it looks like our choice is between 6 x R515's or 6 x R410's depending upon whether we want to allow for the possibility of adding more OSDs at a later date. Yeah, tough call. I would expect that R410s or

Re: [ceph-users] Dell R515 performance and specification question

2013-05-08 Thread Dave Spano
I like the suggestion about the Intel SSDs, but doesn't Dell have a firmware restriction about what drives can go on their controllers? Their cheapest SSD for the 515 $1000+. In a previous conversation, you'd mentioned putting an LSI 9260 in an R515. This would allow someone to get around this r

Re: [ceph-users] Best practice for osd_min_down_reporters

2013-05-08 Thread Wido den Hollander
On 05/07/2013 04:40 PM, Gregory Farnum wrote: On Tuesday, May 7, 2013, Wido den Hollander wrote: Hi, I was just upgrading a 9 nodes, 36 OSD cluster running the next branch from some days ago to the Cuttlefish release. While rebooting the nodes one by one and waiting for a activ

Re: [ceph-users] Dell R515 performance and specification question

2013-05-08 Thread Mark Nelson
On 05/08/2013 09:23 AM, Dave Spano wrote: I like the suggestion about the Intel SSDs, but doesn't Dell have a firmware restriction about what drives can go on their controllers? Their cheapest SSD for the 515 $1000+. In a previous conversation, you'd mentioned putting an LSI 9260 in an R515. This

[ceph-users] CRUSH maps for multiple switches

2013-05-08 Thread Gandalf Corvotempesta
Let's assume 20 OSDs servers and 4x 12 ports switches, 2 for public network and 2 for cluster netowork No link between public switches and no link between cluster switches. first 10 OSD servers connected to public switch1 and the other 10 OSDs connected to public switch2. The same apply for clust

Re: [ceph-users] Mounting CephFS - mount error 5 = Input/output error

2013-05-08 Thread Wyatt Gorman
Does anyone have any ideas about the below authentication error? -- Forwarded message -- From: "Wyatt Gorman" Date: May 7, 2013 1:34 PM Subject: Re: [ceph-users] Mounting CephFS - mount error 5 = Input/output error To: "Jens Kristian Søgaard" , < ceph-users@lists.ceph.com> Here's

[ceph-users] Number of objects per pool?

2013-05-08 Thread Craig Lewis
Is there a practical limit to the number of objects I can store in a pool? I'm planning to use RADOS Gateway, and I'm planning to start by adding about 1M objects to the gateway. Once that initial migration is done and burns in, I want to migrate in another 20M objects. I was planning to use

Re: [ceph-users] HEALTH_WARN after upgrade to cuttlefish

2013-05-08 Thread John Wilkins
James, The output says, " monmap e1: 3 mons at {4= 192.168.200.197:6789/0,7=192.168.200.190:6789/0,8=192.168.200.191:6789/0}, election epoch 1104, quorum 0,1,2 4,7,8" It looks like you have six OSDs (0,1,2,4,7,8) with only 3 OSDs running. The cluster needs a majority. So you'd need 4 of 6 monitor

Re: [ceph-users] HEALTH_WARN after upgrade to cuttlefish

2013-05-08 Thread Sage Weil
On Wed, 8 May 2013, John Wilkins wrote: > James,  > The output says, " monmap e1: 3 mons > at{4=192.168.200.197:6789/0,7=192.168.200.190:6789/0,8=192.168.200.191:6789/0 > }, election epoch 1104, quorum 0,1,2 4,7,8" > > It looks like you have six OSDs (0,1,2,4,7,8) with only 3 OSDs running. The >

Re: [ceph-users] HEALTH_WARN after upgrade to cuttlefish

2013-05-08 Thread Joao Eduardo Luis
On 05/08/2013 10:57 PM, John Wilkins wrote: James, The output says, " monmap e1: 3 mons at {4=192.168.200.197:6789/0,7=192.168.200.190:6789/0,8=192.168.200.191:6789/0 }, election epoch 1104, quorum 0,1,2 4,7,8" It