Re: [ceph-users] Optimal OSD Configuration for 45 drives?

2014-07-28 Thread Christian Balzer
Re-added ML. On Mon, 28 Jul 2014 20:38:37 +1000 Matt Harlum wrote: > > On 27 Jul 2014, at 1:45 am, Christian Balzer wrote: > > > On Sat, 26 Jul 2014 20:49:46 +1000 Matt Harlum wrote: > > > >> > >> On 25 Jul 2014, at 5:54 pm, Christian Balzer wrote: > >> > >>> On Fri, 25 Jul 2014 13:31:34

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Christian Balzer
On Mon, 28 Jul 2014 18:11:33 -0400 Robert Fantini wrote: > "target replication level of 3" > " with a min of 1 across the node level" > > After reading http://ceph.com/docs/master/rados/configuration/ceph-conf/ > , I assume that to accomplish that then set these in ceph.conf ? > > osd pool

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Robert Fantini
"target replication level of 3" " with a min of 1 across the node level" After reading http://ceph.com/docs/master/rados/configuration/ceph-conf/ , I assume that to accomplish that then set these in ceph.conf ? osd pool default size = 3 osd pool default min size = 1 On Mon, Jul 28, 201

Re: [ceph-users] Deployment scenario with 2 hosts

2014-07-28 Thread Michael
You can use multiple "steps" in your crush map in order to do things like choose two different hosts then choose a further OSD on one of the hosts and do another replication so that you can get three replicas onto two hosts without risking ending up with three replicas on a single node. On 28/

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Michael
If you've two rooms then I'd go for two OSD nodes in each room, a target replication level of 3 with a min of 1 across the node level, then have 5 monitors and put the last monitor outside of either room (The other MON's can share with the OSD nodes if needed). Then you've got 'safe' replicatio

[ceph-users] Desktop Ceph Cluster up for grabs!

2014-07-28 Thread Patrick McGarry
Hey cephers, Less than four days left to tweet a photo of "how you're celebrating Ceph's 10th birthday" (be bold, be creative, be awesome) to @Ceph. Include the hastag #cephturns10 and the best photo will win a desktop Ceph cluster built by our own Mark Nelson. https://wiki.ceph.com/Community/Con

Re: [ceph-users] Pool size 2 min_size 1 Advisability?

2014-07-28 Thread Edward Huyer
> >> Running ' ceph osd reweight-by-utilization' clears the issue up > >> temporarily, but additional data inevitably causes certain OSDs to be > >> overloaded again. > >> > > The only time I've ever seen this kind of uneven distribution is when > > using too little (and using the default formula w

Re: [ceph-users] fs as btrfs and ceph journal

2014-07-28 Thread Mark Nelson
Perhaps Cristian is thinking of the clone from journal work that we were talking about last year: http://wiki.ceph.com/Planning/Sideboard/osd%3A_clone_from_journal_on_btrfs I think we never did much beyond Sage's test branch, and it didn't seem to help as much as you would hope. Speaking of wh

Re: [ceph-users] Pool size 2 min_size 1 Advisability?

2014-07-28 Thread Gregory Farnum
On Mon, Jul 28, 2014 at 12:14 PM, Christian Balzer wrote: > On Mon, 28 Jul 2014 14:24:02 + Edward Huyer wrote: > >> > > Ceph has a default pool size of 3. Is it a bad idea to run a pool of >> > > size 2? What about size 2 min_size 1? >> > > >> > min_size 1 is sensible, 2 obviously won't protec

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Robert Fantini
OK for higher availability then 5 nodes is better then 3 . So we'll run 5 . However we want normal operations with just 2 nodes. Is that possible? Eventually 2 nodes will be next building 10 feet away , with a brick wall in between. Connected with Infiniband or better. So one room can go off

Re: [ceph-users] fs as btrfs and ceph journal

2014-07-28 Thread Gregory Farnum
It still helps; the journal does just as much work. Less of the work *can* be in the critical path for IO, but for most of the applications it will be. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sat, Jul 26, 2014 at 2:18 AM, Cristian Falcas wrote: > Hello, > > I'm usin

Re: [ceph-users] slow read speeds from kernel rbd (Firefly 0.80.4)

2014-07-28 Thread Mark Nelson
On 07/28/2014 11:28 AM, Steve Anthony wrote: While searching for more information I happened across the following post (http://dachary.org/?p=2961) which vaguely resembled the symptoms I've been experiencing. I ran tcpdump and noticed what appeared to be a high number of retransmissions on the ho

Re: [ceph-users] Deployment scenario with 2 hosts

2014-07-28 Thread Craig Lewis
That's expected. You need > 50% of the monitors up. If you only have 2 machines, rebooting one means that 50% are up, so the cluster halts operations. That's done on purpose to avoid problems when the cluster is divided in exactly half, and both halves continue to run thinking the other half is

Re: [ceph-users] radosgw monitoring

2014-07-28 Thread Craig Lewis
(Sorry for the duplicate email, I forgot to CC the list) Assuming you're using the default setup (RadosGW, FastCGI, and Apache), it's the same as monitoring a web site. On every node, verify that request for / returns a 200. If the RadosGW agent is down, or FastCGI is mis-configured, the request

[ceph-users] Deployment scenario with 2 hosts

2014-07-28 Thread Don Pinkster
Hi, Currently I am evalutating multiple distributed storage solutions with an S3-like interface. We have two huge machines with big amounts of storage. Is it possible to let these two behave exactly the same with Ceph? My idea is runninng both MON and OSD on these two machines. With quick tests t

Re: [ceph-users] ceph metrics

2014-07-28 Thread Sage Weil
On Mon, 28 Jul 2014, James Eckersall wrote: > Hi, > I'm trying to understand what a lot of the values mean that are reported by > "perf dump" on the ceph admin socket.  I have a collectd plugin which sends > all of these values to graphite. > > Does anyone have a cross-reference list that explains

Re: [ceph-users] Dependency issues in fresh ceph/CentOS 7 install

2014-07-28 Thread Brian Lovett
Simon Ironside writes: > > Hi Brian, > > I have a fresh install working on RHEL 7 running the same version of > python as you. I did have trouble installing from the ceph.com yum repos > though and worked around it by creating and installing from my own local > yum repos instead. > > I then

Re: [ceph-users] slow read speeds from kernel rbd (Firefly 0.80.4)

2014-07-28 Thread Steve Anthony
While searching for more information I happened across the following post (http://dachary.org/?p=2961) which vaguely resembled the symptoms I've been experiencing. I ran tcpdump and noticed what appeared to be a high number of retransmissions on the host where the images are mounted during a read f

[ceph-users] ceph metrics

2014-07-28 Thread James Eckersall
Hi, I'm trying to understand what a lot of the values mean that are reported by "perf dump" on the ceph admin socket. I have a collectd plugin which sends all of these values to graphite. Does anyone have a cross-reference list that explains what they are in more detail? You can glean so much f

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Christian Balzer
On Mon, 28 Jul 2014 11:22:38 +0100 Joao Eduardo Luis wrote: > On 07/28/2014 08:49 AM, Christian Balzer wrote: > > > > Hello, > > > > On Sun, 27 Jul 2014 18:20:43 -0400 Robert Fantini wrote: > > > >> Hello Christian, > >> > >> Let me supply more info and answer some questions. > >> > >> * Our main

Re: [ceph-users] Dependency issues in fresh ceph/CentOS 7 install

2014-07-28 Thread Simon Ironside
Hi Brian, I have a fresh install working on RHEL 7 running the same version of python as you. I did have trouble installing from the ceph.com yum repos though and worked around it by creating and installing from my own local yum repos instead. I then skip the ceph-deploy install step, as I'v

Re: [ceph-users] Pool size 2 min_size 1 Advisability?

2014-07-28 Thread Christian Balzer
On Mon, 28 Jul 2014 14:24:02 + Edward Huyer wrote: > > > Ceph has a default pool size of 3. Is it a bad idea to run a pool of > > > size 2? What about size 2 min_size 1? > > > > > min_size 1 is sensible, 2 obviously won't protect you against dual > > disk failures. Which happen and happen with

Re: [ceph-users] Recommendation to safely avoid problems with osd-failure

2014-07-28 Thread Christian Balzer
See the anti-cephalopod thread on this ML. On Mon, 28 Jul 2014 12:23:02 +0200 Josef Johansson wrote: > Hi, > > I'm trying to compile a strategy to avoid performance problems if osds > or osd hosts fails. > > If I encounter a re-balance of one OSD during mid-day, there'll be > problems with per

[ceph-users] Dependency issues in fresh ceph/CentOS 7 install

2014-07-28 Thread Brian Lovett
I'm installing the latest firefly on a fresh centos 7 machine using the rhel 7 yum repo. I'm getting a few dependency issues when using ceph-deploy install. Mostly it looks like it doesn't like python 2.7. [monitor01][DEBUG ] --> Processing Dependency: libboost_system-mt.so.5() (64bit) for pack

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Sage Weil
On Mon, 28 Jul 2014, Joao Eduardo Luis wrote: > On 07/28/2014 02:07 PM, Robert Fantini wrote: > > Is the '15 minutes or so ' something that can be configured at run time? > > Someone who knows this better than I do should probably chime in, but from a > quick look throughout the code it seems to

Re: [ceph-users] Pool size 2 min_size 1 Advisability?

2014-07-28 Thread Edward Huyer
> > Ceph has a default pool size of 3. Is it a bad idea to run a pool of > > size 2? What about size 2 min_size 1? > > > min_size 1 is sensible, 2 obviously won't protect you against dual disk > failures. > Which happen and happen with near certainty once your cluster gets big > enough. I though

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Joao Eduardo Luis
On 07/28/2014 02:07 PM, Robert Fantini wrote: Is the '15 minutes or so ' something that can be configured at run time? Someone who knows this better than I do should probably chime in, but from a quick look throughout the code it seems to be 'client_mount_interval', which by default is 300 s

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Robert Fantini
Is the '15 minutes or so ' something that can be configured at run time? On Mon, Jul 28, 2014 at 8:44 AM, Joao Eduardo Luis wrote: > (CC'ing ceph-users) > > On 07/28/2014 12:34 PM, Marc wrote: > >> Hi, >> >> >> This said, if out of 3 monitors you have 2 monitors down, your cluster >>> will ce

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Joao Eduardo Luis
(CC'ing ceph-users) On 07/28/2014 12:34 PM, Marc wrote: Hi, This said, if out of 3 monitors you have 2 monitors down, your cluster will cease functioning (no admin commands, no writes or reads served). This is not entirely true. (At least) RBDs will continue being fully functional even if th

[ceph-users] Recommendation to safely avoid problems with osd-failure

2014-07-28 Thread Josef Johansson
Hi, I'm trying to compile a strategy to avoid performance problems if osds or osd hosts fails. If I encounter a re-balance of one OSD during mid-day, there'll be problems with performance right now, if I could see the issue and let it re-balance during evening, that would be great. I.e. if two O

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Joao Eduardo Luis
On 07/28/2014 08:49 AM, Christian Balzer wrote: Hello, On Sun, 27 Jul 2014 18:20:43 -0400 Robert Fantini wrote: Hello Christian, Let me supply more info and answer some questions. * Our main concern is high availability, not speed. Our storage requirements are not huge. However we want good

Re: [ceph-users] Not able to upload object using Horizon(Openstack Dashboard) to Ceph

2014-07-28 Thread Ashish Chandra
Hi Karan, Once you are able to put objects using RadosGW created user name and password 90% of the job is done. We have to follow only these steps afterwards: 1. Put the configurations specified in http://ceph.com/docs/master/radosgw/keystone/ 2. Make sure you create object-service and endpoint i

Re: [ceph-users] OSD weight 0

2014-07-28 Thread Kapil Sharma
It's fixed now. Apparently we can not share a journal across different OSDs. I added a journal /dev/sdc1 (20GB) with my first OSD. I was trying to add the same journal with my second OSD and it was causing the issue. Then I added the secons OSD with a new journal and it worked fine. Thanks, Kapil

[ceph-users] Not able to upload object using Horizon(Openstack Dashboard) to Ceph

2014-07-28 Thread Ashish Chandra
Hi Cephers, I have configured Ceph RadosGW for Swift. I have also set authentication using keystone. Using Swift CLI I can do all stuff viz. uploading container, object, listing. But while using Dashboard I am able to do all the stuff apart from uploading an object. While uploading an object I am

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Robert Fantini
the osd's were created using proxmox web page . there is not data that I want to save.. so I'd like to start from scratch but not do a reinstall of the operating system. I'll check the documentation that you mentioned. On Mon, Jul 28, 2014 at 4:38 AM, Christian Balzer wrote: > > On Mon, 28 J

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Christian Balzer
On Mon, 28 Jul 2014 04:19:16 -0400 Robert Fantini wrote: > I have 3 hosts that i want to use to test new setup... > > Currently they have 3-4 OSD's each. > How did you create the current cluster? ceph-deploy or something withing Proxmox? > Could you suggest a fast way to remove all the OSD's

Re: [ceph-users] Issues compiling Ceph (master branch) on Debian Wheezy (armhf)

2014-07-28 Thread Joao Eduardo Luis
On 07/25/2014 04:54 AM, Deven Phillips wrote: Hi all, I am in the process of installing and setting up Ceph on a group of Allwinner A20 SoC mini computers. They are armhf devices and I have installed Cubian (http://cubian.org/), which is a port of Debian Wheezy. I tried to follow the instru

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Robert Fantini
I have 3 hosts that i want to use to test new setup... Currently they have 3-4 OSD's each. Could you suggest a fast way to remove all the OSD's ? On Mon, Jul 28, 2014 at 3:49 AM, Christian Balzer wrote: > > Hello, > > On Sun, 27 Jul 2014 18:20:43 -0400 Robert Fantini wrote: > > > Hello Chri

Re: [ceph-users] anti-cephalopod question

2014-07-28 Thread Christian Balzer
Hello, On Sun, 27 Jul 2014 18:20:43 -0400 Robert Fantini wrote: > Hello Christian, > > Let me supply more info and answer some questions. > > * Our main concern is high availability, not speed. > Our storage requirements are not huge. > However we want good keyboard response 99.99% of the time

Re: [ceph-users] Could not find module rbd. CentOs 6.4

2014-07-28 Thread Pratik Rupala
Hi Karan, So that means I can't have RBD on 2.6.32. Do you know where can I find source for rbd.ko for other kernel versions like 2.6.34? Regards, Pratik Rupala On 7/28/2014 12:32 PM, Karan Singh wrote: Yes you can use other features like CephFS and Object Store on this kernel release that y

Re: [ceph-users] OSD weight 0

2014-07-28 Thread Karan Singh
Looks like osd.1 has a valid auth ID , which was defined previously. Trust this is your test cluster , try this ceph osd crush rm osd.1 ceph osd rm osd.1 ceph auth del osd.1 Once again try to add osd.1 using ceph-deploy ( prepare and then activate commands ) , check the logs carefully for any

Re: [ceph-users] firefly osds stuck in state booting

2014-07-28 Thread Karan Singh
The output that you have provided says that OSDs are not IN , Try the below ceph osd in osd.0 ceph osd in osd.1 service ceph start osd.0 service ceph start osd.1 If you have 1 more host with 1 disk , add it , starting Ceph Firefly default rep size is 3 - Karan - On 27 Jul 2014, at 11:17, 10

Re: [ceph-users] Could not find module rbd. CentOs 6.4

2014-07-28 Thread Karan Singh
Yes you can use other features like CephFS and Object Store on this kernel release that you are running. - Karan Singh On 28 Jul 2014, at 07:45, Pratik Rupala wrote: > Hi Karan, > > I have basic setup of Ceph storage cluster in active+clean state on Linux > kernel 2.6.32. As per your sugge