Re: [ceph-users] ceph fio read test hangs

2013-07-22 Thread Erwan Velu
On 17/07/2013 14:36, Da Chun wrote: On Ubuntu 13.04, ceph 0.61.4. I was running an fio read test as below, then it hung: root@ceph-node2:/mnt# fio -filename=/dev/rbd1 -direct=1 -iodepth 1 -thread -rw=read -ioengine=psync -bs=4k -size=50G -numjobs=16 -group_reporting -name=mytest mytest: (g=0)

Re: [ceph-users] ceph-deploy mon create doesn't create keyrings

2013-07-22 Thread jose.valeriooropeza
I am using RHEL6. >From the ceph admin machine I executed: ceph-deploy install cephserverX ceph-deploy new cephserverX ceph-deploy mon create cephserverX is there a debug mode more verbose than -v that I can enable, in order to see more what the command ceph-deploy mon create is doing? Thanks

[ceph-users] SSD suggestions as journal

2013-07-22 Thread Gandalf Corvotempesta
I'm looking at some SSDs drives to be used as journal. Seagate 600 should be the better in write intensive operation (like a journal): http://www.storagereview.com/seagate_600_pro_enterprise_ssd_review what do you suggest? Is this good enough ? Should I look for write-intensive operations when se

[ceph-users] Monitor is unable to start after reboot: OSDMonitor::update_from_paxos(bool*) FAILED assert(latest_bl.length() != 0

2013-07-22 Thread peter
Hello, After a reboot one of our monitors is unable to start. We did an upgrade from 0.61.4 to 0.61.5 last week without problems (the monitor restarted just fine). We are getting the following error (I think it is the same as: http://tracker.ceph.com/issues/5704). I might have missed it on t

[ceph-users] adding, deleting or changing privilege for existing cephx users?

2013-07-22 Thread w sun
Does anyone know how to do this or if this is not possible? We try to modify the security scope for an existing cephx user but could not figure out how to add access to a new pool without recreating the user, e.g., ceph auth get-or-create client.svl-ceph-openstack-images mon 'allow r' osd 'allow

Re: [ceph-users] SSD suggestions as journal

2013-07-22 Thread Mark Nelson
On 07/22/2013 06:22 AM, Gandalf Corvotempesta wrote: I'm looking at some SSDs drives to be used as journal. Seagate 600 should be the better in write intensive operation (like a journal): http://www.storagereview.com/seagate_600_pro_enterprise_ssd_review what do you suggest? Is this good enough

Re: [ceph-users] SSD suggestions as journal

2013-07-22 Thread Gandalf Corvotempesta
2013/7/22 Mark Nelson : >> http://www.storagereview.com/seagate_600_pro_enterprise_ssd_review "If you used this SSD for 100% sequential writes, you could theoretically kill it in a little more than a month." very bad. Any other suggestions for SSD device? _

Re: [ceph-users] SSD suggestions as journal

2013-07-22 Thread Mark Nelson
On 07/22/2013 08:16 AM, Gandalf Corvotempesta wrote: 2013/7/22 Mark Nelson : http://www.storagereview.com/seagate_600_pro_enterprise_ssd_review "If you used this SSD for 100% sequential writes, you could theoretically kill it in a little more than a month." very bad. Any other suggestions for

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Chen, Xiaoxi
Hi, My 0.02 : > Secondly, I'm unclear about how OSDs use the journal. It appears they write to the journal (in all cases, can't be turned >off), ack to the client and then read the journal later to write to backing >storage. Is that correct? I would like to say NO, the journal w

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Gandalf Corvotempesta
2013/7/22 Chen, Xiaoxi : > With “journal writeahead”,the data first write to journal ,ack to the > client, and write to OSD, note that, the data always keep in memory before > it write to both OSD and journal,so the write is directly from memory to > OSDs. This mode suite for XFS and EXT4. What ha

Re: [ceph-users] Monitor is unable to start after reboot: OSDMonitor::update_from_paxos(bool*) FAILED assert(latest_bl.length() != 0

2013-07-22 Thread Joao Eduardo Luis
On 07/22/2013 12:33 PM, pe...@2force.nl wrote: Hello, After a reboot one of our monitors is unable to start. We did an upgrade from 0.61.4 to 0.61.5 last week without problems (the monitor restarted just fine). We are getting the following error (I think it is the same as: http://tracker.ceph.c

Re: [ceph-users] adding, deleting or changing privilege for existing cephx users?

2013-07-22 Thread Gregory Farnum
On Mon, Jul 22, 2013 at 5:42 AM, w sun wrote: > Does anyone know how to do this or if this is not possible? We try to modify > the security scope for an existing cephx user but could not figure out how > to add access to a new pool without recreating the user, e.g., > > ceph auth get-or-create cli

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Gandalf Corvotempesta
2013/7/22 Chen, Xiaoxi : > Imaging you have several writes have been flushed to journal and acked,but > not yet write to disk. Now the system crash by kernal panic or power > failure,you will lose your data in ram disk,thus lose data that assumed to be > successful written. The same apply in ca

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Chen, Xiaoxi
发自我的 iPhone 在 2013-7-23,0:21,"Gandalf Corvotempesta" 写道: > 2013/7/22 Chen, Xiaoxi : >> Imaging you have several writes have been flushed to journal and acked,but >> not yet write to disk. Now the system crash by kernal panic or power >> failure,you will lose your data in ram disk,thus lose d

Re: [ceph-users] Monitor is unable to start after reboot: OSDMonitor::update_from_paxos(bool*) FAILED assert(latest_bl.length() != 0

2013-07-22 Thread Joao Eduardo Luis
On 07/22/2013 04:59 PM, pe...@2force.nl wrote: Hi Joao, I have sent you the link to the monitor files. I stopped one other monitor to have a consistent tarball but now it won't start, crashing with the same error message. I hope there is a trick to get it working again because now I only have on

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Jens Kristian Søgaard
Hi, RAM is *MUCH* more reliable than SSD. I've never seen a single RAM module (server grade) failed from the latest 5-6 year. We have had that happen a few times - running primarily IBM hardware. Then you get those nasty MCEs - so I wouldn't run journals in RAM. Unless ofcourse you're only u

Re: [ceph-users] Monitor is unable to start after reboot: OSDMonitor::update_from_paxos(bool*) FAILED assert(latest_bl.length() != 0

2013-07-22 Thread peter
Hi Joao, I have sent you the link to the monitor files. I stopped one other monitor to have a consistent tarball but now it won't start, crashing with the same error message. I hope there is a trick to get it working again because now I only have one monitor working and I don't want to end up

Re: [ceph-users] SSD suggestions as journal

2013-07-22 Thread Chen, Xiaoxi
Basically i think endurance is most important for a ceph journal,since the workload for journal is full write,you can easily caculate how long your ssd will burn out.. even we assume your ssd only run at 100MB/s in average,you will burn out 8TB/day and 240TB/month DCS 3500 is definitely not use

Re: [ceph-users] SSD suggestions as journal

2013-07-22 Thread Wido den Hollander
On 07/22/2013 05:53 PM, Chen, Xiaoxi wrote: > Basically i think endurance is most important for a ceph journal,since the > workload for journal is full write,you can easily caculate how long your ssd > will burn out.. even we assume your ssd only run at 100MB/s in average,you > will burn out 8TB

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Chen, Xiaoxi
发自我的 iPhone 在 2013-7-22,23:16,"Gandalf Corvotempesta" 写道: > 2013/7/22 Chen, Xiaoxi : >> With “journal writeahead”,the data first write to journal ,ack to the >> client, and write to OSD, note that, the data always keep in memory before >> it write to both OSD and journal,so the write is direct

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Mark Nelson
On 07/22/2013 11:26 AM, Chen, Xiaoxi wrote: > > > 发自我的 iPhone > > 在 2013-7-23,0:21,"Gandalf Corvotempesta" 写道: > >> 2013/7/22 Chen, Xiaoxi : >>> Imaging you have several writes have been flushed to journal and acked,but >>> not yet write to disk. Now the system crash by kernal panic or power

Re: [ceph-users] SSD suggestions as journal

2013-07-22 Thread Gandalf Corvotempesta
2013/7/22 Mark Nelson : > I don't have any in my test lab, but the DC S3700 continues to look like a > good option and has a great reputation, but might be a bit pricey. From that > article it looks like the Micron P400m might be worth looking at too, but > seems to be a bit slower. DC S3500 shoul

Re: [ceph-users] SSD suggestions as journal

2013-07-22 Thread Mark Nelson
On 07/22/2013 09:30 AM, Gandalf Corvotempesta wrote: 2013/7/22 Mark Nelson : I don't have any in my test lab, but the DC S3700 continues to look like a good option and has a great reputation, but might be a bit pricey. From that article it looks like the Micron P400m might be worth looking at to

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Oliver Fuckner
Good evening, > I have not yet had the opportunity to try one, but something like the > Marvell Dragonfly might be a very interesting option for servers with > 24+ drives: > > https://origin-www.marvell.com/storage/dragonfly/nvram/ yes, Marvell Dragonfly also looks very promising to me, i like

Re: [ceph-users] ceph & hbase:

2013-07-22 Thread Gregory Farnum
On Fri, Jul 19, 2013 at 11:04 PM, Noah Watkins wrote: > On Fri, Jul 19, 2013 at 8:09 AM, ker can wrote: >> >> With ceph is there any way to influence the data block placement for a >> single file ? > > AFAIK, no... But, this is an interesting twist. New files written out > to HDFS, IIRC, will by

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Mark Nelson
On 07/22/2013 01:02 PM, Oliver Fuckner wrote: > Good evening, > > >> I have not yet had the opportunity to try one, but something like the >> Marvell Dragonfly might be a very interesting option for servers with >> 24+ drives: >> >> https://origin-www.marvell.com/storage/dragonfly/nvram/ > > yes

[ceph-users] Ceph pgs stuck or degraded.

2013-07-22 Thread Gaylord Holder
I have a 12 OSD/3 host set up, and have be stuck with a bunch of stuck pages. I've verified the OSDs are all up and in. The crushmap looks fine. I've tried restarting all the daemons. root@never:/var/lib/ceph/mon# ceph status health HEALTH_WARN 139 pgs degraded; 461 pgs stuck unclean; r

Re: [ceph-users] Ceph pgs stuck or degraded.

2013-07-22 Thread Sage Weil
On Mon, 22 Jul 2013, Gaylord Holder wrote: > > I have a 12 OSD/3 host set up, and have be stuck with a bunch of stuck pages. > > I've verified the OSDs are all up and in. The crushmap looks fine. > I've tried restarting all the daemons. > > > > root@never:/var/lib/ceph/mon# ceph status >h

Re: [ceph-users] adding, deleting or changing privilege for existing cephx users?

2013-07-22 Thread w sun
Ah, that works now with 0.61. I believe we have tried that with 0.56 before and it didn't work. Maybe we have missed something. Anyway, don't have a 0.56 system any more to validate. Thanks much. --weiguo > Date: Mon, 22 Jul 2013 09:23:26 -0700 > From: g...@inktank.com > To: ws...@hotmail.com

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Charles 'Boyo
On Mon, Jul 22, 2013 at 7:10 PM, Mark Nelson wrote: > On 07/22/2013 01:02 PM, Oliver Fuckner wrote: >> Good evening, >> >> >> >> On the second look you see that they use 4 Sandisk X100 SSDs in RAID5 >> and those SSDs only have 80TBytes Write Endurance each... that makes me >> nervous. > > I'm less

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Charles 'Boyo
Hi, On Mon, Jul 22, 2013 at 2:08 AM, Chen, Xiaoxi wrote: > Hi, > > ** ** > > > Can you share any information on the SSD you are using, is it > PCIe connected? > >Depends, if you use HDD as your OSD data disk, a SATA/SAS SSD is > enough for you. Instead of Intel 520, I woul

Re: [ceph-users] Ceph pgs stuck or degraded.

2013-07-22 Thread Gaylord Holder
Sage, The crush tunables did the trick. why? Could you explain what was causing the problem? I've haven't installed 3.9 on my RBD servers yet. Will setting crush tunables back to default or legacy cause me similar problems in the future? Thank you again Sage! -Gaylord On 07/22/2013 02:27

Re: [ceph-users] Ceph pgs stuck or degraded.

2013-07-22 Thread Sage Weil
On Mon, 22 Jul 2013, Gaylord Holder wrote: > Sage, > > The crush tunables did the trick. > > why? Could you explain what was causing the problem? This has a good explanation, I think: http://ceph.com/docs/master/rados/operations/crush-map/#tunables > I've haven't installed 3.9 on my R

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Mikaël Cluseau
On 07/23/13 07:35, Charles 'Boyo wrote: Considering using a mSATA to PCIe adapter with a SATA III mSATA SSD. Any thoughts on what to expect from this combination? Going PCIe I think I would use a SSD card rather than adding yet another (relatively slow) bus. I haven't looked at the models but

Re: [ceph-users] Optimize Ceph cluster (kernel, osd, rbd)

2013-07-22 Thread Dan Mick
I would get the cluster up and running and do some experiments before I spent any time on optimization, much less all this. On 07/20/2013 09:35 AM, Ta Ba Tuan wrote: Please help me! On 07/20/2013 02:11 AM, Ta Ba Tuan wrote: Hi everyone, I have *3 nodes (running MON and MDS)* and *6 data nod

Re: [ceph-users] Ceph pgs stuck or degraded.

2013-07-22 Thread Gaylord Holder
If I understand what the #tunables page is saying, changing the tunables kicks the OSD re-balancing mechanism a bit and resets it to try again. I'll see about getting 3.9 kernel in for my RBD maachines, and reset everything to optimal. Thanks again. -Gaylord On 07/22/2013 04:51 PM, Sage Wei

Re: [ceph-users] Ceph pgs stuck or degraded.

2013-07-22 Thread Sage Weil
On Mon, 22 Jul 2013, Gaylord Holder wrote: > If I understand what the #tunables page is saying, changing the tunables kicks > the OSD re-balancing mechanism a bit and resets it to try again. > > I'll see about getting 3.9 kernel in for my RBD maachines, and reset > everything to optimal. Keep in

Re: [ceph-users] SSD recommendations for OSD journals

2013-07-22 Thread Chen, Xiaoxi
发自我的 iPhone 在 2013-7-23,4:35,"Charles 'Boyo" mailto:charlesb...@gmail.com>> 写道: Hi, On Mon, Jul 22, 2013 at 2:08 AM, Chen, Xiaoxi mailto:xiaoxi.c...@intel.com>> wrote: Hi, > Can you share any information on the SSD you are using, is it PCIe connected? Depends, if you use H

[ceph-users] Openstack on ceph rbd installation failure

2013-07-22 Thread johnu
Hi, I have a three node ceph cluster. ceph -w says health ok . I have openstack in the same cluster and trying to map cinder and glance onto rbd. I have followed steps as given in http://ceph.com/docs/next/rbd/rbd-openstack/ New Settings that is added in cinder.conf for three files volu

[ceph-users] Myanmar Workers for your company

2013-07-22 Thread Oryx Internation
Your email client cannot read this email. To view it online, please go here: http://mailczar247.info/display.php?M=203964&C=6bdb1772177fe8d9135ad2e2dc2b0fe2&S=13&L=8&N=2 To stop receiving these emails:http://mailczar247.info/unsubscribe.php?M=203964&C=6bdb1772177fe8d9135ad2e2dc2b0fe2&L=8&N=13 __

Re: [ceph-users] rgw bucket index

2013-07-22 Thread Yehuda Sadeh
On Sun, Jul 21, 2013 at 12:48 AM, Dominik Mostowiec wrote: > Hi, > Rgw bucket index is in one file (one osd performance issues). > Is there on roudmap sharding or other change to increase performance? > > Excellent question. Note that the upcoming Dupling release will have the ability to separate