[ceph-users] crushmap question

2014-05-13 Thread Cao, Buddy
Hi, I have a crushmap structure likes root->rack->host->osds. I designed the rule below, since I used "chooseleaf...rack" in rule definition, if there is only one rack in the cluster, the ceph gps will always stay at stuck unclean state (that is because the default metadata/data/rbd pool set 2

Re: [ceph-users] crushmap question

2014-05-13 Thread Peter
perhaps group sets of hosts into racks in crushmap. The crushmap doesn't have to strictly map the real world. On 05/13/2014 08:52 AM, Cao, Buddy wrote: Hi, I have a crushmap structure likes root->rack->host->osds. I designed the rule below, since I used "chooseleaf...rack" in rule definition

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Christian Balzer
I'm clearly talking to myself, but whatever. For Greg, I've played with all the pertinent journal and filestore options and TCP nodelay, no changes at all. Is there anybody on this ML who's running a Ceph cluster with a fast network and FAST filestore, so like me with a big HW cache in front of

[ceph-users] Journal SSD durability

2014-05-13 Thread Christian Balzer
Hello, No actual question, just some food for thought and something that later generations can scour from the ML archive. I'm planning another Ceph storage cluster, this time a "classic" Ceph design, 3 storage nodes with 8 HDDs for OSDs and 4 SSDs for OS and journal. When juggling the budget fo

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Alexandre DERUMIER
Hi Christian, I'm going to test a full ssd cluster in coming months, I'll send result on the mailing. Do you have tried to use 1 osd by physical disk ? (without raid6) Maybe they are bottleneck in osd daemon, and using osd daemon by disk could help. - Mail original - De: "Christi

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Christian Balzer
Hello, On Tue, 13 May 2014 11:33:27 +0200 (CEST) Alexandre DERUMIER wrote: > Hi Christian, > > I'm going to test a full ssd cluster in coming months, > I'll send result on the mailing. > Looking forward to that. > > Do you have tried to use 1 osd by physical disk ? (without raid6) > No, if y

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Mark Kirkwood
On thing that would put me off the 530 is lack on power off safety (capacitor or similar). Given the job of the journal, I think an SSD that has some guarantee of write integrity is crucial - so yeah the DC3500 or DC3700 seem like the best choices. Regards Mark On 13/05/14 21:31, Christian B

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Xabier Elkano
El 13/05/14 11:31, Christian Balzer escribió: > Hello, > > No actual question, just some food for thought and something that later > generations can scour from the ML archive. > > I'm planning another Ceph storage cluster, this time a "classic" Ceph > design, 3 storage nodes with 8 HDDs for OSDs an

[ceph-users] Fwd: What is link and unlink options used for in radosgw-admin

2014-05-13 Thread Wenjun Huang
Begin forwarded message: > From: Wenjun Huang > Subject: What is link and unlink options used for in radosgw-admin > Date: May 13, 2014 at 2:55:18 PM GMT+8 > To: ceph-us...@ceph.com > > Hello, everyone > > I am now confused with the options of link & unlink in radosgw-admin utility. > > In m

Re: [ceph-users] Bulk storage use case

2014-05-13 Thread Cédric Lemarchand
Thanks for your answers Craig, it seems this is a niche use case for Ceph, not a lot of replies from the ML. Cheers -- Cédric Lemarchand > Le 11 mai 2014 à 00:35, Craig Lewis a écrit : > >> On 5/10/14 12:43 , Cédric Lemarchand wrote: >> Hi Craig, >> >> Thanks, I really appreciate the well d

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Alexandre DERUMIER
>>It might, but at the IOPS I'm seeing anybody using SSD for file storage >>should have screamed out already. >>Also given the CPU usage I'm seeing during that test run such a setup >>would probably require 32+ cores. Just found this: https://objects.dreamhost.com/inktankweb/Inktank_Hardware_

Re: [ceph-users] Bulk storage use case

2014-05-13 Thread Dan van der Ster
Hi, I think you're not getting many replies simply because those are rather large servers and not many have such hardware in prod. We run with 24x3TB drives, 64GB ram, one 10Gbit NIC. Memory-wise there are no problems. Throughput-wise, the bottleneck is somewhere between the NIC (~1GB/s) and

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Christian Balzer
On Tue, 13 May 2014 22:03:11 +1200 Mark Kirkwood wrote: > On thing that would put me off the 530 is lack on power off safety > (capacitor or similar). Given the job of the journal, I think an SSD > that has some guarantee of write integrity is crucial - so yeah the > DC3500 or DC3700 seem like

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Christian Balzer
On Tue, 13 May 2014 12:07:12 +0200 Xabier Elkano wrote: > El 13/05/14 11:31, Christian Balzer escribió: > > Hello, > > > > No actual question, just some food for thought and something that later > > generations can scour from the ML archive. > > > > I'm planning another Ceph storage cluster, this

[ceph-users] Rados GW Method not allowed

2014-05-13 Thread Georg Höllrigl
Hello, System Ubuntu 14.04 Ceph 0.80 I'm getting either a 405 Method Not Allowed or a 403 Permission Denied from Radosgw. Here is what I get from radosgw: HTTP/1.1 405 Method Not Allowed Date: Tue, 13 May 2014 12:21:43 GMT Server: Apache Accept-Ranges: bytes Content-Length: 82 Content-Type:

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Christian Balzer
Hello, On Tue, 13 May 2014 13:36:49 +0200 (CEST) Alexandre DERUMIER wrote: > >>It might, but at the IOPS I'm seeing anybody using SSD for file > >>storage should have screamed out already. > >>Also given the CPU usage I'm seeing during that test run such a setup > >>would probably require 32+

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Xabier Elkano
El 13/05/14 14:23, Christian Balzer escribió: > On Tue, 13 May 2014 12:07:12 +0200 Xabier Elkano wrote: > >> El 13/05/14 11:31, Christian Balzer escribió: >>> Hello, >>> >>> No actual question, just some food for thought and something that later >>> generations can scour from the ML archive. >>> >>

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Christian Balzer
On Tue, 13 May 2014 14:46:23 +0200 Xabier Elkano wrote: > El 13/05/14 14:23, Christian Balzer escribió: > > On Tue, 13 May 2014 12:07:12 +0200 Xabier Elkano wrote: > > > >> El 13/05/14 11:31, Christian Balzer escribió: > >>> Hello, > >>> > >>> No actual question, just some food for thought and som

[ceph-users] Performance stats

2014-05-13 Thread yalla.gnan.kumar
Hi All, Is there a way by which we can measure the performance of Ceph block devices ? (Example : I/O stats, data to identify bottlenecks etc). Also what are the available ways in which we can compare Ceph storage performance with other storage solutions ? Thanks Kumar _

Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-13 Thread Gilles Mocellin
Le 12/05/2014 15:45, Uwe Grohnwaldt a écrit : Hi, yes, we use it in production. I can stop/kill the tgt on one server and XenServer goes to the second one. We enabled multipathing in xenserver. In our setup we haven't multiple ip-ranges so we scan/login the second target on xenserverstartup w

[ceph-users] Monitoring ceph statistics

2014-05-13 Thread Adrian Banasiak
Hi, i am working with test Ceph cluster and now I want to implement Zabbix monitoring with items such as: - whoe cluster IO (for example ceph -s -> recovery io 143 MB/s, 35 objects/s) - pg statistics I would like to create single script in python to retrive values using rados python module, but t

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Alexandre DERUMIER
>>For what it's worth, my cluster gives me 4100 IOPS with the sequential fio >>run below and 7200 when doing random reads (go figure). Of course I made >>sure these came come the pagecache of the storage nodes, no disk I/O >>reported at all and the CPUs used just 1 core per OSD. >>--- >>fio --

Re: [ceph-users] Migrate whole clusters

2014-05-13 Thread Kyle Bader
> Anyway replacing set of monitors means downtime for every client, so > I`m in doubt if 'no outage' word is still applicable there. Taking the entire quorum down for migration would be bad. It's better to add one in the new location, remove one at the old, ad infinitum. -- Kyle ___

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Christian Balzer
On Tue, 13 May 2014 16:09:28 +0200 (CEST) Alexandre DERUMIER wrote: > >>For what it's worth, my cluster gives me 4100 IOPS with the sequential > >>fio run below and 7200 when doing random reads (go figure). Of course > >>I made sure these came come the pagecache of the storage nodes, no > >>disk I

Re: [ceph-users] NFS over CEPH - best practice

2014-05-13 Thread Andrei Mikhailovsky
Dima, do you have any examples / howtos for this? I would love to give it a go. Cheers - Original Message - From: "Dimitri Maziuk" To: ceph-users@lists.ceph.com Sent: Monday, 12 May, 2014 3:38:11 PM Subject: Re: [ceph-users] NFS over CEPH - best practice On 5/12/2014 4:52 AM, Andr

Re: [ceph-users] Ceph with VMWare / XenServer

2014-05-13 Thread Andrei Mikhailovsky
Uwe, do you mind sharing your storage and xenserver iscsi config files? Also, what is your performance like? Thanks - Original Message - From: "Uwe Grohnwaldt" To: ceph-users@lists.ceph.com Sent: Monday, 12 May, 2014 2:45:43 PM Subject: Re: [ceph-users] Ceph with VMWare / XenServ

[ceph-users] Monitoring ceph statistics using rados python module

2014-05-13 Thread Adrian Banasiak
Hi, i am working with test Ceph cluster and now I want to implement Zabbix monitoring with items such as: - whoe cluster IO (for example ceph -s -> recovery io 143 MB/s, 35 objects/s) - pg statistics I would like to create single script in python to retrive values using rados python module, but t

Re: [ceph-users] NFS over CEPH - best practice

2014-05-13 Thread Dimitri Maziuk
On 5/13/2014 9:43 AM, Andrei Mikhailovsky wrote: Dima, do you have any examples / howtos for this? I would love to give it a go. Not really: I haven't done this myself. Google for "tgtd failover with heartbeat", you should find something useful. The setups I have are heartbeat (3.0.x) managi

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Alexandre DERUMIER
>>Actually check your random read output again, you gave it the wrong >>parameter, it needs to be randread, not rand-read. oops, sorry. I got around 7500iops with randread. >>Your cluster isn't that old (the CPUs are in the same ballpark) Yes, this is 6-7 year old server. (this xeons were release

Re: [ceph-users] Monitoring ceph statistics using rados python module

2014-05-13 Thread Haomai Wang
Not sure your demand. I use "ceph --admin-daemon /var/run/ceph/ceph-osd.x.asok perf dump" to get the monitor infos. And the result can be parsed by simplejson easily via python. On Tue, May 13, 2014 at 10:56 PM, Adrian Banasiak wrote: > Hi, i am working with test Ceph cluster and now I want to i

[ceph-users] Lost access to radosgw after crash?

2014-05-13 Thread Brian Rak
I hit a "bug" where radosgw crashed with -101> 2014-05-13 15:26:07.188494 7fde82886820 0 ERROR: FCGX_Accept_r returned -24 0> 2014-05-13 15:26:07.193772 7fde82886820 -1 rgw/rgw_main.cc: In function 'virtual void RGWProcess::RGWWQ::_clear()' thread 7fde82886820 time 2014-05-13 15:26:07.1

Re: [ceph-users] ceph firefly PGs in active+clean+scrubbing state

2014-05-13 Thread Fabrizio G. Ventola
I've upgraded to 0.80.1 on a testing instance: the cluster gets cyclically active+clean+deep scrubbing for a little while and then reaches active+clean status. I'm not worried about this, I think it's normal, but I didn't have this behaviour on emperor 0.72.2. Cheers, Fabrizio On 13 May 2014 06:0

[ceph-users] Occasional Missing Admin Sockets

2014-05-13 Thread Mike Dawson
All, I have a recurring issue where the admin sockets (/var/run/ceph/ceph-*.*.asok) may vanish on a running cluster while the daemons keep running (or restart without my knowledge). I see this issue on a dev cluster running Ubuntu and Ceph Emperor/Firefly, deployed with ceph-deploy using Upst

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Alexandre DERUMIER
I have just done some test, with fio-rbd, (http://telekomcloud.github.io/ceph/2014/02/26/ceph-performance-analysis_fio_rbd.html) directly from the kvm host,(not from the vm). 1 fio job: around 8000iops 2 differents parralel fio job (on different rbd volume) : around 8000iops by fio job ! cpu

Re: [ceph-users] Lost access to radosgw after crash?

2014-05-13 Thread Yehuda Sadeh
On Tue, May 13, 2014 at 8:52 AM, Brian Rak wrote: > I hit a "bug" where radosgw crashed with > > -101> 2014-05-13 15:26:07.188494 7fde82886820 0 ERROR: FCGX_Accept_r > returned -24 too many files opened. You probably need to adjust your limits. > > 0> 2014-05-13 15:26:07.193772 7fde828868

Re: [ceph-users] Lost access to radosgw after crash?

2014-05-13 Thread Brian Rak
I upgraded to 0.80.1 to see if that helped. It didn't change anything, but I'm now seeing more useful errors: 2014-05-13 16:27:32.954007 7f5183cfc700 0 RGWGC::process() failed to acquire lock on gc.10 2014-05-13 16:27:48.098428 7f5183cfc700 0 RGWGC::process() failed to acquire lock on gc.14

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Christian Balzer
On Tue, 13 May 2014 18:10:25 +0200 (CEST) Alexandre DERUMIER wrote: > I have just done some test, > > with fio-rbd, > (http://telekomcloud.github.io/ceph/2014/02/26/ceph-performance-analysis_fio_rbd.html) > > directly from the kvm host,(not from the vm). > > > 1 fio job: around 8000iops > 2 d

Re: [ceph-users] Monitoring ceph statistics using rados python module

2014-05-13 Thread Adrian Banasiak
Thanks for sugestion with admin daemon but it looks like single osd oriented. I have used perf dump on mon socket and it output some interesting data in case of monitoring whole cluster: { "cluster": { "num_mon": 4, "num_mon_quorum": 4, "num_osd": 29, "num_osd_up": 29, "num_

Re: [ceph-users] Lost access to radosgw after crash?

2014-05-13 Thread Brian Rak
On 5/13/2014 12:29 PM, Yehuda Sadeh wrote: On Tue, May 13, 2014 at 8:52 AM, Brian Rak wrote: I hit a "bug" where radosgw crashed with -101> 2014-05-13 15:26:07.188494 7fde82886820 0 ERROR: FCGX_Accept_r returned -24 too many files opened. You probably need to adjust your limits. 0>

Re: [ceph-users] Slow IOPS on RBD compared to journal and backing devices

2014-05-13 Thread Alexandre DERUMIER
>>You didn't specify what you did, but i assume you did read test. yes, indeed >>Those scale, as in running fio in multiple VMs in parallel gives me about >>6200 IOPS each, so much better than the 7200 for a single one. >>And yes, the client CPU is quite busy. oh ok ! >>However my real, orig

Re: [ceph-users] Lost access to radosgw after crash?

2014-05-13 Thread Brian Rak
This turns out to have been a configuration change to nginx that I forgot I had made. It wasn't passing all the http options through any more, so authentication was failing. On 5/13/2014 12:43 PM, Brian Rak wrote: On 5/13/2014 12:29 PM, Yehuda Sadeh wrote: On Tue, May 13, 2014 at 8:52 AM, B

Re: [ceph-users] Monitoring ceph statistics using rados python module

2014-05-13 Thread Mike Dawson
Adrian, Yes, it is single OSD oriented. Like Haomai, we monitor perf dumps from individual OSD admin sockets. On new enough versions of ceph, you can do 'ceph daemon osd.x perf dump', which is a shorter way to ask for the same output as 'ceph --admin-daemon /var/run/ceph/ceph-osd.x.asok perf

Re: [ceph-users] Monitoring ceph statistics using rados python module

2014-05-13 Thread Don Talton (dotalton)
python-cephclient may be of some use to you https://github.com/dmsimard/python-cephclient > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Mike Dawson > Sent: Tuesday, May 13, 2014 10:04 AM > To: Adrian Banasiak; Haomai Wang > Cc: ceph-us

Re: [ceph-users] too slowly upload on ceph object storage

2014-05-13 Thread Stephen Taylor
By way of follow-up, I have done quite a bit more additional testing here and my problem seems to be exclusive to rbd-fuse on Ubuntu 14.04. So probably not related to this thread I chimed in on last week. I am able to get read and write speeds in the hundreds of megabytes per second with this s

Re: [ceph-users] Where is the SDK of ceph object storage

2014-05-13 Thread Gregory Farnum
On Mon, May 12, 2014 at 11:55 PM, wsnote wrote: > Hi, everyone! > Where can I find the SDK of ceph object storage? > Python: boto > C++: libs3 which I found in the src of ceph and github.com/ceph/libs3. > where are that of other language? Does ceph supply them? > Otherwise I use the SDK of Amazon

Re: [ceph-users] crushmap question

2014-05-13 Thread Gregory Farnum
You just use a type other than "rack" in your chooseleaf rule. In your case, "host". When using chooseleaf, the bucket type you specify is the failure domain which it must segregate across. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Tue, May 13, 2014 at 12:52 AM, Cao, B

Re: [ceph-users] Occasional Missing Admin Sockets

2014-05-13 Thread Gregory Farnum
On Tue, May 13, 2014 at 9:06 AM, Mike Dawson wrote: > All, > > I have a recurring issue where the admin sockets > (/var/run/ceph/ceph-*.*.asok) may vanish on a running cluster while the > daemons keep running Hmm. >(or restart without my knowledge). I'm guessing this might be involved: > I see

Re: [ceph-users] Migrate whole clusters

2014-05-13 Thread Fred Yang
I have to say I'm shocked to see the suggestion is rbd import/export if 'you care the data'. These kind of operation is common use case and should be an essential part of any distributed storage. What if I have a hundred node cluster running for years and need to do hardware refresh? There are no c

Re: [ceph-users] Occasional Missing Admin Sockets

2014-05-13 Thread Mike Dawson
Greg/Loic, I can confirm that "logrotate --force /etc/logrotate.d/ceph" removes the monitor admin socket on my boxes running 0.80.1 just like the description in Issue 7188 [0]. 0: http://tracker.ceph.com/issues/7188 Should that bug be reopened? Thanks, Mike Dawson On 5/13/2014 2:10 PM, Gr

Re: [ceph-users] Migrate whole clusters

2014-05-13 Thread Gregory Farnum
On Tue, May 13, 2014 at 11:36 AM, Fred Yang wrote: > I have to say I'm shocked to see the suggestion is rbd import/export if 'you > care the data'. These kind of operation is common use case and should be an > essential part of any distributed storage. What if I have a hundred node > cluster runni

Re: [ceph-users] Occasional Missing Admin Sockets

2014-05-13 Thread Gregory Farnum
Yeah, I just did so. :( -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Tue, May 13, 2014 at 11:41 AM, Mike Dawson wrote: > Greg/Loic, > > I can confirm that "logrotate --force /etc/logrotate.d/ceph" removes the > monitor admin socket on my boxes running 0.80.1 just like th

Re: [ceph-users] ceph firefly PGs in active+clean+scrubbing state

2014-05-13 Thread Michael
Anyone still seeing this issue on 0.80.1 you'll probable need to dump out your scrub list "ceph pg dump | grep scrub" then find the OSD listed as the acting primary for the PG stuck scrubbing and stop it a bit more aggressively. I found that the acting primary for a PG stuck in scrub status was

[ceph-users] Ceph 0.80.1 delete/recreate data/metadata pools

2014-05-13 Thread Michael
Hi All, Seems commit 2adc534a72cc199c8b11dbdf436258cbe147101b has removed the ability to delete and recreate the data and metadata pools using osd pool delete (Returns Error EBUSY - Is in use by CephFS). Currently have no mds running as I'm no longer using CephFS and so it's not mounted anywh

Re: [ceph-users] Ceph 0.80.1 delete/recreate data/metadata pools

2014-05-13 Thread Michael
Answered my own question. Created two new pools, used mds newfs on them and then deleted the original pools and renamed the new ones. -Michael On 13/05/2014 22:20, Michael wrote: Hi All, Seems commit 2adc534a72cc199c8b11dbdf436258cbe147101b has removed the ability to delete and recreate the

Re: [ceph-users] Migrate whole clusters

2014-05-13 Thread Gandalf Corvotempesta
2014-05-13 21:21 GMT+02:00 Gregory Farnum : > You misunderstand. Migrating between machines for incrementally > upgrading your hardware is normal behavior and well-tested (likewise > for swapping in all-new hardware, as long as you understand the IO > requirements involved). So is decommissioning o

Re: [ceph-users] Migrate whole clusters

2014-05-13 Thread Gregory Farnum
Assuming you have the spare throughput-/IOPS for Ceph to do its thing without disturbing your clients, this will work fine. -Greg On Tuesday, May 13, 2014, Gandalf Corvotempesta < gandalf.corvotempe...@gmail.com> wrote: > 2014-05-13 21:21 GMT+02:00 Gregory Farnum > >: > > You misunderstand. Migr

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Craig Lewis
On 5/13/14 05:15 , Christian Balzer wrote: On Tue, 13 May 2014 22:03:11 +1200 Mark Kirkwood wrote: On thing that would put me off the 530 is lack on power off safety (capacitor or similar). Given the job of the journal, I think an SSD that has some guarantee of write integrity is crucial - so y

Re: [ceph-users] Rados GW Method not allowed

2014-05-13 Thread Andrei Mikhailovsky
Georg, I've had similar issues when I had a "+" character in my secret key. Not all clients support it. You might need to escape this with \ and see if it works. Andrei - Original Message - From: "Georg Höllrigl" To: ceph-users@lists.ceph.com Sent: Tuesday, 13 May, 2014 1:30:14

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Kyle Bader
> TL;DR: Power outages are more common than your colo facility will admit. Seconded. I've seen power failures in at least 4 different facilities and all of them had the usual gamut of batteries/generators/etc. Some of those facilities I've seen problems multiple times in a single year. Even a data

Re: [ceph-users] Journal SSD durability

2014-05-13 Thread Dimitri Maziuk
On 05/13/2014 05:49 PM, Kyle Bader wrote: >> TL;DR: Power outages are more common than your colo facility will admit. > > Seconded. I've seen power failures in at least 4 different facilities > and all of them had the usual gamut of batteries/generators/etc. Some > of those facilities I've seen pr

Re: [ceph-users] Monitoring ceph statistics using rados python module

2014-05-13 Thread Craig Lewis
On 5/13/14 09:33 , Adrian Banasiak wrote: Thanks for sugestion with admin daemon but it looks like single osd oriented. I have used perf dump on mon socket and it output some interesting data in case of monitoring whole cluster: { "cluster": { "num_mon": 4, "num_mon_quorum": 4, "num

Re: [ceph-users] Monitoring ceph statistics using rados python module

2014-05-13 Thread Kai Zhang
Hi Adrian, You may be interested in "rados -p poo_name df --format json", although it's pool oriented, you could probably add the values together :) Regards, Kai 在 2014-05-13 08:33:11,"Adrian Banasiak" 写道: Thanks for sugestion with admin daemon but it looks like single osd oriented. I have u

[ceph-users] Error while initializing OSD directory

2014-05-13 Thread Srinivasa Rao Ragolu
Hi All, I am following manual steps to create osd node. while executing below command, i am facing error like below #ceph-osd -i 1 --mkfs --mkkey 2014-05-14 05:04:12.097585 7f91c99007c0 -1 ** ERROR: error creating empty object store in /var/lib/ceph/osd/-: (2) No such file or directory But dir

Re: [ceph-users] crushmap question

2014-05-13 Thread Cao, Buddy
Thanks Gregory so much,it solved the problem! Wei Cao (Buddy) -Original Message- From: Gregory Farnum [mailto:g...@inktank.com] Sent: Wednesday, May 14, 2014 2:00 AM To: Cao, Buddy Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] crushmap question You just use a type other than

Re: [ceph-users] crushmap question

2014-05-13 Thread Cao, Buddy
BTW, I'd like to know, after I change the "from rack" to "from host", if I add more racks with host/osds in the cluster, will ceph choose the osds for pg only from one zone? or ceph will randomly choose from several different zones? Wei Cao (Buddy) -Original Message- From: Cao, Buddy