i use 0.87, in side ceph.conf, do not have mds.0 related config
i did
*root@ceph06-vm:~# ceph mds rm 0 mds.ceph06-vm*
*mds gid 0 dne*
2015-01-05 11:15 GMT+07:00 Lindsay Mathieson :
> Did you remove the mds.0 entry from ceph.conf?
>
> On 5 January 2015 at 14:13, debian Only wrote:
>
>> i h
Some low level caching might help, flashcache, dmcache,etc…
But that may hurt the reliability to some extent , and make it harder for
operator ☺
From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of
Lindsay Mathieson
Sent: Monday, January 5, 2015 12:14 PM
To: Christian Balzer
Hi Sage,
On Tue, Dec 23, 2014 at 10:10 PM, Sage Weil wrote:
>
> This fun issue came up again in the form of 10422:
>
> http://tracker.ceph.com/issues/10422
>
> I think we have 3 main options:
>
> 1. Ask users to do a mon scrub prior to upgrade to
> ensure it is safe. If a mon is out of s
Hi Udo,
Lindsay did this for performance reasons so that the data is spread evenly
over the disks, I believe it has been accepted that the remaining 2tb on the
3tb disks will not be used.
Nick
-Original Message-
From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Ud
Hi!
I have a cluster with 106 osds and disk usage is varying from 166gb to
316gb. Disk usage is highly correlated to number of pg per osd (no surprise
here). Is there a reason for ceph to allocate more pg on some nodes?
The biggest osds are 30, 42 and 69 (300gb+ each) and the smallest are 87,
33
On 05-01-15 11:04, ivan babrou wrote:
Hi!
I have a cluster with 106 osds and disk usage is varying from 166gb to
316gb. Disk usage is highly correlated to number of pg per osd (no
surprise here). Is there a reason for ceph to allocate more pg on some
nodes?
The biggest osds are 30, 42 and 69
HI all,
i'm digging in the src code of ceph0.80.6, i think the following code in void
Monitor::timecheck_start_round() in moitor.cc from line 3160 to 3168 is very
strange,
in my opinion, the red highligted part should becurr_time -
timecheck_round_start < max ,
that is if time elapsed l
I don't use it, but there is also a RBD kernel driver, to use RBD
devices directly, without qemu.
http://ceph.com/docs/master/rbd/rbd-ko/
--
Thomas Lemarchand
Cloud Solutions SAS - Responsable des systèmes d'information
On sam., 2015-01-03 at 08:11 +1000, Lindsay Mathieson wrote:
> On Fri, 2 J
On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote:
> Hi!
>
> I have a cluster with 106 osds and disk usage is varying from 166gb to
> 316gb. Disk usage is highly correlated to number of pg per osd (no
> surprise here). Is there a reason for ceph to allocate more pg on some
> nodes?
>
In essenc
On Tue, Dec 30, 2014 at 10:42 AM, Lindsay Mathieson
wrote:
> $ ceph mds newfs metadata data --yes-i-really-mean-it
*Please* don't advise people to run this without also telling them
that it will erase their filesystem!
The original poster was asking how to delete an MDS so that he could
move it
I try myself, use these command and recover to Health Ok now. But i do
not know why were these command work, in my opinion , fail mds node first
and rm failed mds node
root@ceph01-vm:~#* ceph mds fail 0*
failed mds.0
root@ceph01-vm:~# ceph -s
cluster 075f1aae-48de-412e-b024-b0f014dbc8cf
FYI Zheng pushed a fix for this to master recently
(https://github.com/ceph/ceph/pull/3286), it's currently pending
backport for the previous release series.
In the meantime, you can safely ignore this message.
Cheers,
John
On Tue, Dec 30, 2014 at 12:30 PM, Jiri Kanicky wrote:
> Hi.
>
> I have
On 5 January 2015 at 14:20, Christian Balzer wrote:
> On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote:
>
> > Hi!
> >
> > I have a cluster with 106 osds and disk usage is varying from 166gb to
> > 316gb. Disk usage is highly correlated to number of pg per osd (no
> > surprise here). Is there a
Hi
I have stragne problem when I try to start radosgw on docker container.
When I have single container with 1 radosgw process inside everyting
is ok, I have good performance i think, for single test thread: 80
put/s for 4k objects and radosgw with debug enabled.
When I start two containers on the
On Mon, 5 Jan 2015 09:21:16 AM Nick Fisk wrote:
> Lindsay did this for performance reasons so that the data is spread evenly
> over the disks, I believe it has been accepted that the remaining 2tb on the
> 3tb disks will not be used.
Exactly, thanks Nick.
I only have a terabyte of data, and its n
On Mon, Jan 5, 2015 at 10:56 AM, debian Only wrote:
> I try myself, use these command and recover to Health Ok now. But i do not
> know why were these command work, in my opinion , fail mds node first and rm
> failed mds node
>
> root@ceph01-vm:~# ceph mds fail 0
Okay, good -- this is telling c
On 01/05/2015 01:39 PM, ivan babrou wrote:
> On 5 January 2015 at 14:20, Christian Balzer wrote:
>
>> On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote:
>>
>>> Hi!
>>>
>>> I have a cluster with 106 osds and disk usage is varying from 166gb to
>>> 316gb. Disk usage is highly correlated to number
I've been having good results with OMD (Check_MK + Nagios)
There is a plugin for Ceph as well that I made a small modification to, to
work with a wider range of cluster sizes
http://www.spinics.net/lists/ceph-users/msg13355.html
Nick
-Original Message-
From: ceph-users [mailto:ceph-user
On Fri, Jan 2, 2015 at 9:59 PM, Gregory Farnum wrote:
> I think it's just for service isolation that people recommend splitting
> them. The only technical issue I can think of is that you don't want to put
> kernel clients on the same OS as an OSD (due to deadlock scenarios under
> memory pressure
On Sunday, January 4, 2015, Chen, Xiaoxi wrote:
> You could use rbd info to see the block_name_prefix, the
> object name consist like ., so for
> example, rb.0.ff53.3d1b58ba.e6ad should be the th object of
> the volume with block_name_prefix rb.0.ff53.3d1b58ba.
>
> $ rbd info hug
On Mon, 05 Jan 2015 13:53:56 +0100 Wido den Hollander wrote:
> On 01/05/2015 01:39 PM, ivan babrou wrote:
> > On 5 January 2015 at 14:20, Christian Balzer wrote:
> >
> >> On Mon, 5 Jan 2015 14:04:28 +0400 ivan babrou wrote:
> >>
> >>> Hi!
> >>>
> >>> I have a cluster with 106 osds and disk usage
Hi all,
I think I have a subtle problem with either understanding CRUSH or in
the actual implementation of my CRUSH map.
Consider the following CRUSH map: http://paste.debian.net/hidden/085b3f20/
I have 3 chassis' with 7 nodes each (6 of them OSDs). Size is 3,
min_size is 2 on all pools.
If i re
On Mon, 5 Jan 2015 01:15:03 PM Nick Fisk wrote:
> I've been having good results with OMD (Check_MK + Nagios)
>
> There is a plugin for Ceph as well that I made a small modification to, to
> work with a wider range of cluster sizes
Thanks, I'll check it out.
Currently trying zabbix, seems more s
Can you post the output of 'ceph pg dump'?
-Sam
On Mon, Jan 5, 2015 at 6:16 AM, Christopher Kunz wrote:
> Hi all,
>
> I think I have a subtle problem with either understanding CRUSH or in
> the actual implementation of my CRUSH map.
>
> Consider the following CRUSH map: http://paste.debian.net/hi
Hi All,
Would anybody have an idea a) If it's possible and b) if it's a good idea
to have more EC chunks than the total number of hosts?
For instance if I wanted to have a k=6 m=2, but only across 4 hosts and I
wanted to be able to withstand 1 host failure and 1 disk failure(any host),
wo
On Sat, Nov 22, 2014 at 12:47 AM, Vinod H I wrote:
> Thanks for the clarification.
> Now I have done exactly as you suggested.
> "us-east" is the master zone and "us-west" is the secondary zone.
> Each zone has two system users "us-east" and "us-west".
> These system users have same access/secret
Hey cephers,
I know it's hardly the time that one might think about summer, but the
org submissions for this year's Google Summer of Code (GSoC) are fast
approaching. Last year we had some great idea submissions and a couple
of great student participants:
http://ceph.com/gsoc2014/
If you have an
Hi Nick,
What about subdividing your hosts using containers ? For instance four
container per host on your four hosts which gives you 16 hosts. When you add
more hosts you move containers around and reduce the number of containers per
host. But you don't need to change the rulesets.
Cheers
On
Hello Yehuda,
It is exactly the same:
2015-01-05 17:39:06.303832 7fd4458ed780 20 rados->read obj-ofs=0 read_ofs=0
read_len=524288
{ "user_id": "johndoe3",
"display_name": "John Doe_3",
"email": "joh...@example.com",
"suspended": 0,
"max_buckets": 1000,
"auid": 0,
"subusers": [
One more thing Yehuda,
In radosgw log in Slave Zone:
2015-01-05 17:22:42.188108 7fe4b66d2780 20 enqueued request req=0xbc1f50
2015-01-05 17:22:42.188125 7fe4b66d2780 20 RGWWQ:
2015-01-05 17:22:42.188126 7fe4b66d2780 20 req: 0xbc1f50
2015-01-05 17:22:42.188129 7fe4b66d2780 10 allocated request
Hello,
I have a requirement to aggregate the result from various RadosGW and then
put across the result to the end user, is it possible with RadosGW.
Lets assume I have an object: o1, under bucket b1 for user u1 in Master
Zone of US Region(u1b1o1)
Metadata for US region is replicated to MasterZon
Hi,
I'm trying to build Ceph on my RHEL (Scientific Linux 7 - Nitrogen), with
3.10.0.
I am using the configure script and I am now stuck on "libkeyutils not found".
I can't seem to find the right library for this. What Is the right yum update
name for this library?
Any help appreciated.
Thanks
Pa
On 01/05/2015 11:14 AM, x...@csvcn.cn wrote:
HI all,
i'm digging in the src code of ceph0.80.6, i think the following code in
void Monitor::timecheck_start_round() in moitor.cc from line 3160 to
3168 is very strange,
in my opinion, the red highligted part should be *curr_time -
timecheck_round
Rebalancing is almost finished, but things got even worse:
http://i.imgur.com/0HOPZil.png
Moreover, one pg is in active+remapped+wait_backfill+backfill_toofull state:
2015-01-05 19:39:31.995665 mon.0 [INF] pgmap v3979616: 5832 pgs: 23
active+remapped+wait_backfill, 1
active+remapped+wait_backfill
It took me a while to figure out the callout script since it wasn't
documented anywhere easy. This is what I wrote down, it could be helpful to
you or others:
1.
Add the hook script to the ceph.conf file of each OSD
osd crush location hook = /path/to/script
1.
Install the script a
On 01/05/2015 11:26 AM, Garg, Pankaj wrote:
> I’m trying to build Ceph on my RHEL (Scientific Linux 7 – Nitrogen),
> with 3.10.0.
>
> I am using the configure script and I am now stuck on “libkeyutils not
> found”.
>
> I can’t seem to find the right library for this. What Is the right yum
> updat
Hi Ken,
Spot-on. After much googling I just figured out the name and yes it is very
un-intuitive named keyutils-libs-devel.
And yes the names for debian etc are "libkeyutils-dev".
I'm not a linux expert and this stuff does drive me crazy.
Thanks
Pankaj
-Original Message-
From: Ken Dreye
Hi,
I am wondering how a PG gets marked backfill_toofull.
I reweighted several OSDs using ceph osd crush reweight. As expected,
PG began moving around (backfilling).
Some PGs got marked +backfilling (~10), some +wait_backfill (~100).
But some are marked +backfill_toofull. My OSDs are betwee
If Ceph snapshots work like VM snapshots (and I don't have any reason to
believe otherwise), the snapshot will never grow larger than the size of
the base image. If the same blocks are rewritten, then they are just
rewritten in the snapshot and don't take any extra space. The snapshot
functions dif
Hi Ceph Users,
We've got a Ceph cluster we've built, and we're experiencing issues with slow
or hung IO's, even running 'rados bench' on the OSD cluster. Things start out
great, ~600 MB/s, then rapidly drops off as the test waits for IO's. Nothing
seems to be taxed... the system just seems to
On Mon, Jan 5, 2015 at 12:11 PM, Robert LeBlanc wrote:
> If Ceph snapshots work like VM snapshots (and I don't have any reason to
> believe otherwise), the snapshot will never grow larger than the size of the
> base image. If the same blocks are rewritten, then they are just rewritten
> in the sna
Hi,
Yeah, the performance when the system isn't waiting for slow IO's is definitely
acceptable for what I'm doing, it's just the handful of slow IO's messing up
the overall latency.
Bill
From: Gonzalo Aguilar Delgado [gagui...@aguilardelgado.com]
Sent: Monday, J
Thanks for the reply.
14 and 18 happened to show up during that run, but its certainly not only those
OSD's. It seems to vary each run. Just from the runs I've done today I've
seen the following pairs of OSD's:
['0,13', '0,18', '0,24', '0,25', '0,32', '0,34', '0,36', '10,22', '11,30',
'12,28
Hi BIll,
From your log excerpt, it looks like your slow requests are happening on OSDs
14 and 18. Is it always these two OSDs?
If you don't have a long recovery time (e.g., the cluster is just full of test
data), maybe you could try setting OSDs 14 and 18 out and re-benching?
Alternatively I s
On Sat, Jan 3, 2015 at 8:53 PM, Christian Balzer wrote:
> On Sat, 3 Jan 2015 16:21:29 +1000 Lindsay Mathieson wrote:
>
>> I just added 4 OSD's to my 2 OSD "cluster" (2 Nodes, now have 3 OSD's per
>> node).
>>
>> Given its the weekend and not in use, I've set them all to weight 1, but
>> looks like
45 matches
Mail list logo