Re: [ceph-users] How to remove mds from cluster

2015-01-05 Thread debian Only
i use 0.87, in side ceph.conf, do not have mds.0 related config i did *root@ceph06-vm:~# ceph mds rm 0 mds.ceph06-vm* *mds gid 0 dne* 2015-01-05 11:15 GMT+07:00 Lindsay Mathieson : > Did you remove the mds.0 entry from ceph.conf? > > On 5 January 2015 at 14:13, debian Only wrote: > >> i h

Re: [ceph-users] Worthwhile setting up Cache tier with small leftover SSD partions?

2015-01-05 Thread Chen, Xiaoxi
Some low level caching might help, flashcache, dmcache,etc… But that may hurt the reliability to some extent , and make it harder for operator ☺ From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Lindsay Mathieson Sent: Monday, January 5, 2015 12:14 PM To: Christian Balzer

Re: [ceph-users] full osdmaps in mon txns

2015-01-05 Thread Dan van der Ster
Hi Sage, On Tue, Dec 23, 2014 at 10:10 PM, Sage Weil wrote: > > This fun issue came up again in the form of 10422: > > http://tracker.ceph.com/issues/10422 > > I think we have 3 main options: > > 1. Ask users to do a mon scrub prior to upgrade to > ensure it is safe. If a mon is out of s

Re: [ceph-users] Improving Performance with more OSD's?

2015-01-05 Thread Nick Fisk
Hi Udo, Lindsay did this for performance reasons so that the data is spread evenly over the disks, I believe it has been accepted that the remaining 2tb on the 3tb disks will not be used. Nick -Original Message- From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Ud

[ceph-users] Different disk usage on different OSDs

2015-01-05 Thread ivan babrou
Hi! I have a cluster with 106 osds and disk usage is varying from 166gb to 316gb. Disk usage is highly correlated to number of pg per osd (no surprise here). Is there a reason for ceph to allocate more pg on some nodes? The biggest osds are 30, 42 and 69 (300gb+ each) and the smallest are 87, 33

Re: [ceph-users] Different disk usage on different OSDs

2015-01-05 Thread Wido den Hollander
On 05-01-15 11:04, ivan babrou wrote: Hi! I have a cluster with 106 osds and disk usage is varying from 166gb to 316gb. Disk usage is highly correlated to number of pg per osd (no surprise here). Is there a reason for ceph to allocate more pg on some nodes? The biggest osds are 30, 42 and 69

[ceph-users] ceph timecheck bug on monitors

2015-01-05 Thread x...@csvcn.cn
HI all, i'm digging in the src code of ceph0.80.6, i think the following code in void Monitor::timecheck_start_round() in moitor.cc from line 3160 to 3168 is very strange, in my opinion, the red highligted part should becurr_time - timecheck_round_start < max , that is if time elapsed l

Re: [ceph-users] Not running multiple services on the same machine?

2015-01-05 Thread Thomas Lemarchand
I don't use it, but there is also a RBD kernel driver, to use RBD devices directly, without qemu. http://ceph.com/docs/master/rbd/rbd-ko/ -- Thomas Lemarchand Cloud Solutions SAS - Responsable des systèmes d'information On sam., 2015-01-03 at 08:11 +1000, Lindsay Mathieson wrote: > On Fri, 2 J

Re: [ceph-users] Different disk usage on different OSDs

2015-01-05 Thread Christian Balzer
On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote: > Hi! > > I have a cluster with 106 osds and disk usage is varying from 166gb to > 316gb. Disk usage is highly correlated to number of pg per osd (no > surprise here). Is there a reason for ceph to allocate more pg on some > nodes? > In essenc

Re: [ceph-users] How to remove mds from cluster

2015-01-05 Thread John Spray
On Tue, Dec 30, 2014 at 10:42 AM, Lindsay Mathieson wrote: > $ ceph mds newfs metadata data --yes-i-really-mean-it *Please* don't advise people to run this without also telling them that it will erase their filesystem! The original poster was asking how to delete an MDS so that he could move it

Re: [ceph-users] How to remove mds from cluster

2015-01-05 Thread debian Only
I try myself, use these command and recover to Health Ok now. But i do not know why were these command work, in my opinion , fail mds node first and rm failed mds node root@ceph01-vm:~#* ceph mds fail 0* failed mds.0 root@ceph01-vm:~# ceph -s cluster 075f1aae-48de-412e-b024-b0f014dbc8cf

Re: [ceph-users] cephfs kernel module reports error on mount

2015-01-05 Thread John Spray
FYI Zheng pushed a fix for this to master recently (https://github.com/ceph/ceph/pull/3286), it's currently pending backport for the previous release series. In the meantime, you can safely ignore this message. Cheers, John On Tue, Dec 30, 2014 at 12:30 PM, Jiri Kanicky wrote: > Hi. > > I have

Re: [ceph-users] Different disk usage on different OSDs

2015-01-05 Thread ivan babrou
On 5 January 2015 at 14:20, Christian Balzer wrote: > On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote: > > > Hi! > > > > I have a cluster with 106 osds and disk usage is varying from 166gb to > > 316gb. Disk usage is highly correlated to number of pg per osd (no > > surprise here). Is there a

[ceph-users] radosgw on docker container - high CPU usage even on idle state

2015-01-05 Thread Dominik Mostowiec
Hi I have stragne problem when I try to start radosgw on docker container. When I have single container with 1 radosgw process inside everyting is ok, I have good performance i think, for single test thread: 80 put/s for 4k objects and radosgw with debug enabled. When I start two containers on the

Re: [ceph-users] Improving Performance with more OSD's?

2015-01-05 Thread Lindsay Mathieson
On Mon, 5 Jan 2015 09:21:16 AM Nick Fisk wrote: > Lindsay did this for performance reasons so that the data is spread evenly > over the disks, I believe it has been accepted that the remaining 2tb on the > 3tb disks will not be used. Exactly, thanks Nick. I only have a terabyte of data, and its n

Re: [ceph-users] How to remove mds from cluster

2015-01-05 Thread John Spray
On Mon, Jan 5, 2015 at 10:56 AM, debian Only wrote: > I try myself, use these command and recover to Health Ok now. But i do not > know why were these command work, in my opinion , fail mds node first and rm > failed mds node > > root@ceph01-vm:~# ceph mds fail 0 Okay, good -- this is telling c

Re: [ceph-users] Different disk usage on different OSDs

2015-01-05 Thread Wido den Hollander
On 01/05/2015 01:39 PM, ivan babrou wrote: > On 5 January 2015 at 14:20, Christian Balzer wrote: > >> On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote: >> >>> Hi! >>> >>> I have a cluster with 106 osds and disk usage is varying from 166gb to >>> 316gb. Disk usage is highly correlated to number

Re: [ceph-users] Improving Performance with more OSD's?

2015-01-05 Thread Nick Fisk
I've been having good results with OMD (Check_MK + Nagios) There is a plugin for Ceph as well that I made a small modification to, to work with a wider range of cluster sizes http://www.spinics.net/lists/ceph-users/msg13355.html Nick -Original Message- From: ceph-users [mailto:ceph-user

Re: [ceph-users] Not running multiple services on the same machine?

2015-01-05 Thread John Spray
On Fri, Jan 2, 2015 at 9:59 PM, Gregory Farnum wrote: > I think it's just for service isolation that people recommend splitting > them. The only technical issue I can think of is that you don't want to put > kernel clients on the same OS as an OSD (due to deadlock scenarios under > memory pressure

Re: [ceph-users] rbd resize (shrink) taking forever and a day

2015-01-05 Thread Jake Young
On Sunday, January 4, 2015, Chen, Xiaoxi wrote: > You could use rbd info to see the block_name_prefix, the > object name consist like ., so for > example, rb.0.ff53.3d1b58ba.e6ad should be the th object of > the volume with block_name_prefix rb.0.ff53.3d1b58ba. > > $ rbd info hug

Re: [ceph-users] Different disk usage on different OSDs

2015-01-05 Thread Christian Balzer
On Mon, 05 Jan 2015 13:53:56 +0100 Wido den Hollander wrote: > On 01/05/2015 01:39 PM, ivan babrou wrote: > > On 5 January 2015 at 14:20, Christian Balzer wrote: > > > >> On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote: > >> > >>> Hi! > >>> > >>> I have a cluster with 106 osds and disk usage

[ceph-users] CRUSH question - failing to rebalance after failure test

2015-01-05 Thread Christopher Kunz
Hi all, I think I have a subtle problem with either understanding CRUSH or in the actual implementation of my CRUSH map. Consider the following CRUSH map: http://paste.debian.net/hidden/085b3f20/ I have 3 chassis' with 7 nodes each (6 of them OSDs). Size is 3, min_size is 2 on all pools. If i re

Re: [ceph-users] Improving Performance with more OSD's?

2015-01-05 Thread Lindsay Mathieson
On Mon, 5 Jan 2015 01:15:03 PM Nick Fisk wrote: > I've been having good results with OMD (Check_MK + Nagios) > > There is a plugin for Ceph as well that I made a small modification to, to > work with a wider range of cluster sizes Thanks, I'll check it out. Currently trying zabbix, seems more s

Re: [ceph-users] CRUSH question - failing to rebalance after failure test

2015-01-05 Thread Samuel Just
Can you post the output of 'ceph pg dump'? -Sam On Mon, Jan 5, 2015 at 6:16 AM, Christopher Kunz wrote: > Hi all, > > I think I have a subtle problem with either understanding CRUSH or in > the actual implementation of my CRUSH map. > > Consider the following CRUSH map: http://paste.debian.net/hi

[ceph-users] Erasure Encoding Chunks > Number of Hosts

2015-01-05 Thread Nick Fisk
Hi All, Would anybody have an idea a) If it's possible and b) if it's a good idea to have more EC chunks than the total number of hosts? For instance if I wanted to have a k=6 m=2, but only across 4 hosts and I wanted to be able to withstand 1 host failure and 1 disk failure(any host), wo

Re: [ceph-users] Regarding Federated Gateways - Zone Sync Issues

2015-01-05 Thread Yehuda Sadeh
On Sat, Nov 22, 2014 at 12:47 AM, Vinod H I wrote: > Thanks for the clarification. > Now I have done exactly as you suggested. > "us-east" is the master zone and "us-west" is the secondary zone. > Each zone has two system users "us-east" and "us-west". > These system users have same access/secret

[ceph-users] Google Summer of Code Prep Begins!

2015-01-05 Thread Patrick McGarry
Hey cephers, I know it's hardly the time that one might think about summer, but the org submissions for this year's Google Summer of Code (GSoC) are fast approaching. Last year we had some great idea submissions and a couple of great student participants: http://ceph.com/gsoc2014/ If you have an

Re: [ceph-users] Erasure Encoding Chunks > Number of Hosts

2015-01-05 Thread Loic Dachary
Hi Nick, What about subdividing your hosts using containers ? For instance four container per host on your four hosts which gives you 16 hosts. When you add more hosts you move containers around and reduce the number of containers per host. But you don't need to change the rulesets. Cheers On

Re: [ceph-users] Regarding Federated Gateways - Zone Sync Issues

2015-01-05 Thread hemant burman
Hello Yehuda, It is exactly the same: 2015-01-05 17:39:06.303832 7fd4458ed780 20 rados->read obj-ofs=0 read_ofs=0 read_len=524288 { "user_id": "johndoe3", "display_name": "John Doe_3", "email": "joh...@example.com", "suspended": 0, "max_buckets": 1000, "auid": 0, "subusers": [

Re: [ceph-users] Regarding Federated Gateways - Zone Sync Issues

2015-01-05 Thread hemant burman
One more thing Yehuda, In radosgw log in Slave Zone: 2015-01-05 17:22:42.188108 7fe4b66d2780 20 enqueued request req=0xbc1f50 2015-01-05 17:22:42.188125 7fe4b66d2780 20 RGWWQ: 2015-01-05 17:22:42.188126 7fe4b66d2780 20 req: 0xbc1f50 2015-01-05 17:22:42.188129 7fe4b66d2780 10 allocated request

[ceph-users] Aggregate Results from Multiple RadosGW

2015-01-05 Thread hemant burman
Hello, I have a requirement to aggregate the result from various RadosGW and then put across the result to the end user, is it possible with RadosGW. Lets assume I have an object: o1, under bucket b1 for user u1 in Master Zone of US Region(u1b1o1) Metadata for US region is replicated to MasterZon

[ceph-users] Building Ceph

2015-01-05 Thread Garg, Pankaj
Hi, I'm trying to build Ceph on my RHEL (Scientific Linux 7 - Nitrogen), with 3.10.0. I am using the configure script and I am now stuck on "libkeyutils not found". I can't seem to find the right library for this. What Is the right yum update name for this library? Any help appreciated. Thanks Pa

Re: [ceph-users] ceph timecheck bug on monitors

2015-01-05 Thread Joao Eduardo Luis
On 01/05/2015 11:14 AM, x...@csvcn.cn wrote: HI all, i'm digging in the src code of ceph0.80.6, i think the following code in void Monitor::timecheck_start_round() in moitor.cc from line 3160 to 3168 is very strange, in my opinion, the red highligted part should be *curr_time - timecheck_round

Re: [ceph-users] Different disk usage on different OSDs

2015-01-05 Thread ivan babrou
Rebalancing is almost finished, but things got even worse: http://i.imgur.com/0HOPZil.png Moreover, one pg is in active+remapped+wait_backfill+backfill_toofull state: 2015-01-05 19:39:31.995665 mon.0 [INF] pgmap v3979616: 5832 pgs: 23 active+remapped+wait_backfill, 1 active+remapped+wait_backfill

Re: [ceph-users] Crush Map and SSD Pools

2015-01-05 Thread Robert LeBlanc
It took me a while to figure out the callout script since it wasn't documented anywhere easy. This is what I wrote down, it could be helpful to you or others: 1. Add the hook script to the ceph.conf file of each OSD osd crush location hook = /path/to/script 1. Install the script a

Re: [ceph-users] Building Ceph

2015-01-05 Thread Ken Dreyer
On 01/05/2015 11:26 AM, Garg, Pankaj wrote: > I’m trying to build Ceph on my RHEL (Scientific Linux 7 – Nitrogen), > with 3.10.0. > > I am using the configure script and I am now stuck on “libkeyutils not > found”. > > I can’t seem to find the right library for this. What Is the right yum > updat

Re: [ceph-users] Building Ceph

2015-01-05 Thread Garg, Pankaj
Hi Ken, Spot-on. After much googling I just figured out the name and yes it is very un-intuitive named keyutils-libs-devel. And yes the names for debian etc are "libkeyutils-dev". I'm not a linux expert and this stuff does drive me crazy. Thanks Pankaj -Original Message- From: Ken Dreye

[ceph-users] backfill_toofull, but OSDs not full

2015-01-05 Thread c3
Hi, I am wondering how a PG gets marked backfill_toofull. I reweighted several OSDs using ceph osd crush reweight. As expected, PG began moving around (backfilling). Some PGs got marked +backfilling (~10), some +wait_backfill (~100). But some are marked +backfill_toofull. My OSDs are betwee

Re: [ceph-users] rbd snapshot slow restore

2015-01-05 Thread Robert LeBlanc
If Ceph snapshots work like VM snapshots (and I don't have any reason to believe otherwise), the snapshot will never grow larger than the size of the base image. If the same blocks are rewritten, then they are just rewritten in the snapshot and don't take any extra space. The snapshot functions dif

[ceph-users] Slow/Hung IOs

2015-01-05 Thread Sanders, Bill
Hi Ceph Users, We've got a Ceph cluster we've built, and we're experiencing issues with slow or hung IO's, even running 'rados bench' on the OSD cluster. Things start out great, ~600 MB/s, then rapidly drops off as the test waits for IO's. Nothing seems to be taxed... the system just seems to

Re: [ceph-users] rbd snapshot slow restore

2015-01-05 Thread Gregory Farnum
On Mon, Jan 5, 2015 at 12:11 PM, Robert LeBlanc wrote: > If Ceph snapshots work like VM snapshots (and I don't have any reason to > believe otherwise), the snapshot will never grow larger than the size of the > base image. If the same blocks are rewritten, then they are just rewritten > in the sna

Re: [ceph-users] Slow/Hung IOs

2015-01-05 Thread Sanders, Bill
Hi, Yeah, the performance when the system isn't waiting for slow IO's is definitely acceptable for what I'm doing, it's just the handful of slow IO's messing up the overall latency. Bill From: Gonzalo Aguilar Delgado [gagui...@aguilardelgado.com] Sent: Monday, J

Re: [ceph-users] Slow/Hung IOs

2015-01-05 Thread Sanders, Bill
Thanks for the reply. 14 and 18 happened to show up during that run, but its certainly not only those OSD's. It seems to vary each run. Just from the runs I've done today I've seen the following pairs of OSD's: ['0,13', '0,18', '0,24', '0,25', '0,32', '0,34', '0,36', '10,22', '11,30', '12,28

Re: [ceph-users] Slow/Hung IOs

2015-01-05 Thread Lincoln Bryant
Hi BIll, From your log excerpt, it looks like your slow requests are happening on OSDs 14 and 18. Is it always these two OSDs? If you don't have a long recovery time (e.g., the cluster is just full of test data), maybe you could try setting OSDs 14 and 18 out and re-benching? Alternatively I s

Re: [ceph-users] Added OSD's, weighting

2015-01-05 Thread Gregory Farnum
On Sat, Jan 3, 2015 at 8:53 PM, Christian Balzer wrote: > On Sat, 3 Jan 2015 16:21:29 +1000 Lindsay Mathieson wrote: > >> I just added 4 OSD's to my 2 OSD "cluster" (2 Nodes, now have 3 OSD's per >> node). >> >> Given its the weekend and not in use, I've set them all to weight 1, but >> looks like