Re: [ceph-users] Issue while building Jewel on ARM

2016-06-20 Thread Gregory Farnum
On Mon, Jun 20, 2016 at 5:28 AM, Daleep Singh Bais wrote: > Dear All, > > I am getting below error message while trying to build Jewel on ARM. Any > help / suggestion will be appreciated. > > g++: error: unrecognized command line option '-momit-leaf-frame-pointer' > g++: error: unrecognized comman

Re: [ceph-users] cluster down during backfilling, Jewel tunables and client IO optimisations

2016-06-20 Thread Gregory Farnum
On Mon, Jun 20, 2016 at 8:33 AM, Daniel Swarbrick wrote: > We have just updated our third cluster from Infernalis to Jewel, and are > experiencing similar issues. > > We run a number of KVM virtual machines (qemu 2.5) with RBD images, and > have seen a lot of D-state processes and even jbd/2 timeo

Re: [ceph-users] cephfs snapshots

2016-06-22 Thread Gregory Farnum
On Wednesday, June 22, 2016, Kenneth Waegeman wrote: > Hi all, > > In Jewel ceph fs snapshots are still experimental. Does someone has a clue > when this would become stable, or how experimental this is ? > We're not sure yet. Probably it will follow stable multi-MDS; we're thinking about redoin

[ceph-users] cephfs snapshots

2016-06-22 Thread Gregory Farnum
. Are > any of the know issues with snapshots issues that result in the loss of > non-snapshot data or a cluster? > On Jun 22, 2016 2:16 PM, "Gregory Farnum" wrote: > >> On Wednesday, June 22, 2016, Kenneth Waegeman >> wrote: >> >>> Hi all, >

Re: [ceph-users] osd current.remove.me.somenumber

2016-06-27 Thread Gregory Farnum
On Sat, Jun 25, 2016 at 11:22 AM, Mike Miller wrote: > Hi, > > what is the meaning of the directory "current.remove.me.846930886" is > /var/lib/ceph/osd/ceph-14? If you're using btrfs, I believe that's a no-longer-required snapshot of the current state of the system. If you're not, I've no idea w

Re: [ceph-users] Expected behavior of blacklisted host and cephfs

2016-06-30 Thread Gregory Farnum
On Thu, Jun 30, 2016 at 9:09 AM, Mauricio Garavaglia wrote: > Hello, > > What's the expected behavior of a host that has a cephfs mounted and is then > blacklisted? It doesn't seem to fail in a consistent way. Thanks Well, once blacklisted it won't be allowed to make updates to the OSDs or MDS. D

Re: [ceph-users] object size changing after a pg repair

2016-06-30 Thread Gregory Farnum
On Wed, Jun 29, 2016 at 10:50 PM, Goncalo Borges wrote: > Hi Shinobu > >> Sorry probably I don't understand your question properly. >> Is what you're worry about that object mapped to specific pg could be >> overwritten on different osds? > > Not really. I was worried by seeing object sizes chang

Re: [ceph-users] Improving metadata throughput

2016-06-30 Thread Gregory Farnum
On Wed, Jun 29, 2016 at 2:02 PM, Daniel Davidson wrote: > I am starting to work with and benchmark our ceph cluster. While throughput > is so far looking good, metadata performance so far looks to be suffering. > Is there anything that can be done to speed up the response time of looking > throug

Re: [ceph-users] mds standby + standby-reply upgrade

2016-06-30 Thread Gregory Farnum
On Thu, Jun 30, 2016 at 1:03 PM, Dzianis Kahanovich wrote: > Upgraded infernalis->jewel (git, Gentoo). Upgrade passed over global > stop/restart everything oneshot. > > Infernalis: e5165: 1/1/1 up {0=c=up:active}, 1 up:standby-replay, 1 up:standby > > Now after upgrade start and next mon restart,

Re: [ceph-users] Active MON aborts on Jewel 10.2.2 with FAILED assert(info.state == MDSMap::STATE_STANDBY

2016-07-05 Thread Gregory Farnum
Thanks for the report; created a ticket and somebody will get on it shortly. http://tracker.ceph.com/issues/16592 -Greg On Sun, Jul 3, 2016 at 5:55 PM, Bill Sharer wrote: > I was working on a rolling upgrade on Gentoo to Jewel 10.2.2 from 10.2.0. > However now I can't get a monitor quorum going

Re: [ceph-users] mds standby + standby-reply upgrade

2016-07-05 Thread Gregory Farnum
On Mon, Jul 4, 2016 at 12:38 PM, Dzianis Kahanovich wrote: > Gregory Farnum пишет: >> On Thu, Jun 30, 2016 at 1:03 PM, Dzianis Kahanovich wrote: >>> Upgraded infernalis->jewel (git, Gentoo). Upgrade passed over global >>> stop/restart everything oneshot. >>&g

Re: [ceph-users] what's the meaning of 'removed_snaps' of `ceph osd pool ls detail`?

2016-07-07 Thread Gregory Farnum
On Thu, Jul 7, 2016 at 1:07 AM, 秀才 wrote: > Hi,All:) > > i have made a cache-tier, > but i do not know message 'removed_snaps > [1~1,3~6,b~6,13~c,21~4,26~1,28~1a,4e~4,53~5,5c~5,63~1,65~4,6b~4]'. > i have not snapped any thing yet. When you take snapshots, it generally creates a lot of tracking da

Re: [ceph-users] Using two roots for the same pool

2016-07-11 Thread Gregory Farnum
I'm not looking at the docs, but I think you need an "emit" statement after every choose. -Greg On Monday, July 11, 2016, George Shuklin wrote: > Hello. > > I want to try CRUSH rule with following idea: > take one OSD from root with SSD drives (and use it as primary). > take two OSD from root wi

Re: [ceph-users] ceph-fuse segfaults ( jewel 10.2.2)

2016-07-11 Thread Gregory Farnum
Oh, is this one of your custom-built packages? Are they using tcmalloc? That difference between VSZ and RSS looks like a glibc malloc problem. -Greg On Mon, Jul 11, 2016 at 12:04 AM, Goncalo Borges wrote: > Hi John... > > Thank you for replying. > > Here is the result of the tests you asked but I

Re: [ceph-users] Using two roots for the same pool

2016-07-11 Thread Gregory Farnum
uleset 1 > type replicated > min_size 1 > max_size 4 > step take ssd > step choose firstn 2 type room > step chooseleaf firstn 2 type host > step emit > } > > Bob > > On Mon, Jul 11, 2016 at 9:19 AM, Gregory Farnum wrote: >> >> I'm not looking at the docs,

Re: [ceph-users] cephfs change metadata pool?

2016-07-12 Thread Gregory Farnum
I'm not at all sure that rados cppool actually captures everything (it might). Doug has been working on some similar stuff for disaster recovery testing and can probably walk you through moving over. But just how large *is* your metadata pool in relation to others? Having a too-large pool doesn't

Re: [ceph-users] cephfs-journal-tool lead to data missing and show up

2016-07-15 Thread Gregory Farnum
On Thu, Jul 14, 2016 at 1:42 AM, txm wrote: > I am a user of cephfs. > > Recently i met a problem by using the cephfs-journal-tool. > > There were some strange things happened below. > > 1.After use the cephfs-journal-tool and cephfs-table-tool(i came up with the > "negative object nums” issues,

Re: [ceph-users] CephFS | Recursive stats not displaying with GNU ls

2016-07-18 Thread Gregory Farnum
Specifically, this has caused trouble with certain (common?) rsync configurations. -Greg On Monday, July 18, 2016, David wrote: > Thanks Zheng, I should have checked that. > > Sean, from the commit: > > When rbytes mount option is enabled, directory size is recursive size. > Recursive size is no

Re: [ceph-users] ceph admin socket from non root

2016-07-18 Thread Gregory Farnum
I'm not familiar with how it's set up but skimming and searching through the code I'm not seeing anything, no. We've got a chown but no chmod. That's a reasonably feature idea though, and presumably you could add a chmod it to your init scripts? -Greg On Mon, Jul 18, 2016 at 3:02 AM, Stefan Priebe

Re: [ceph-users] CephFS write performance

2016-07-19 Thread Gregory Farnum
On Tue, Jul 19, 2016 at 9:39 AM, Patrick Donnelly wrote: > On Tue, Jul 19, 2016 at 10:25 AM, Fabiano de O. Lucchese > wrote: >> I configured the cluster to replicate data twice (3 copies), so these >> numbers fall within my expectations. So far so good, but here's comes the >> issue: I configured

Re: [ceph-users] CephFS write performance

2016-07-21 Thread Gregory Farnum
n in write throughput to around 170 Mbps. The system > took about 5-6 seconds before it started writing any files and was > constantly at 150 – 180 Mbps write through put when the directory was > mounted using FUSE. > > Any additional thoughts? Would the problem be due to my NFS clien

Re: [ceph-users] Uncompactable Monitor Store at 69GB -- Re: Cluster in warn state, not sure what to do next.

2016-07-21 Thread Gregory Farnum
On Thu, Jul 21, 2016 at 11:54 AM, Salwasser, Zac wrote: > Rephrasing for brevity – I have a monitor store that is 69GB and won’t > compact any further on restart or with ‘tell compact’. Has anyone dealt > with this before? The monitor can't trim OSD maps over a period where PGs are unclean; you'

[ceph-users] CephFS snapshot preferred behaviors

2016-07-25 Thread Gregory Farnum
All, I spent several days last week examining our current snapshot implementation and thinking about how it could be improved. As part of that ongoing effort, I'd love to know what user expectations are about behavior. (I'm going to open up a ceph-devel thread on the implementation details shortly,

Re: [ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Gregory Farnum
On Mon, Jul 25, 2016 at 7:38 PM, Oliver Dzombic wrote: > Hi, > > currently some productive stuff is down, because it can not be accessed > through cephfs. > > Client server restart, did not help. > Cluster restart, did not help. > > Only ONE directory inside cephfs has this issue. > > All other di

Re: [ceph-users] cephfs failed to rdlock, waiting

2016-07-25 Thread Gregory Farnum
4M 85.10 1.72 288 > 11 0.21799 1.0 223G 193G 31155M 86.38 1.75 285 > TOTAL 53926G 26654G 27272G 49.43 > MIN/MAX VAR: 0.72/1.84 STDDEV: 21.46 > > > -- > Mit freundlichen Gruessen / Best regards > > Oliver Dzombic > IP-Interactive > > mailt

Re: [ceph-users] cephfs - mds hardware recommendation for 40 million files and 500 users

2016-07-26 Thread Gregory Farnum
On Tue, Jul 26, 2016 at 1:53 PM, Mike Miller wrote: > Hi, > > we have started to migrate user homes to cephfs with the mds server 32GB > RAM. With multiple rsync threads copying this seems to be undersized; the > mds process consumes all memory 32GB fitting about 4 million caps. > > Any hardware r

Re: [ceph-users] Searchable metadata and objects in Ceph

2016-07-27 Thread Gregory Farnum
On Wed, Jul 27, 2016 at 9:17 AM, Andrey Ptashnik wrote: > Hello team, > > We are looking for ways to store metadata with objects and make this metadata > searchable. > For example if we store an image of the car in Ceph we would like to be able > to attach metadata like model, make, year, damage

Re: [ceph-users] CephFS snapshot preferred behaviors

2016-07-27 Thread Gregory Farnum
On Wed, Jul 27, 2016 at 2:51 PM, Patrick Donnelly wrote: > On Mon, Jul 25, 2016 at 5:41 PM, Gregory Farnum wrote: >> Some specific questions: >> * Right now, we allow users to rename snapshots. (This is newish, so >> you may not be aware of it if you've been using sna

Re: [ceph-users] ceph-fuse (jewel 10.2.2): No such file or directory issues

2016-07-27 Thread Gregory Farnum
On Wed, Jul 27, 2016 at 6:13 PM, Goncalo Borges wrote: > Dear cephfsers :-) > > We saw some weirdness in cephfs that we do not understand. > > We were helping some user which complained that her batch system job outputs > were not produced in cephfs. > > Please note that we are using ceph-fuse (je

Re: [ceph-users] ceph-fuse (jewel 10.2.2): No such file or directory issues

2016-07-28 Thread Gregory Farnum
On Wed, Jul 27, 2016 at 6:37 PM, Goncalo Borges wrote: > Hi Greg > > Thanks for replying. Answer inline. > > > >>> Dear cephfsers :-) >>> >>> We saw some weirdness in cephfs that we do not understand. >>> >>> We were helping some user which complained that her batch system job >>> outputs >>> were

Re: [ceph-users] CephFS snapshot preferred behaviors

2016-07-29 Thread Gregory Farnum
On Thu, Jul 28, 2016 at 2:44 PM, Alexandre Oliva wrote: > On Jul 25, 2016, Gregory Farnum wrote: > >> * Right now, we allow users to rename snapshots. (This is newish, so >> you may not be aware of it if you've been using snapshots for a >> while.) Is that an importa

Re: [ceph-users] Tunables Jewel - request for clarification

2016-08-01 Thread Gregory Farnum
On Sun, Jul 31, 2016 at 6:53 PM, Goncalo Borges wrote: > Dear cephers. > > I would like to request some clarification on migrating from legacy to > optimal (jewel) tunables. > > We have recently migrated from infernalis to Jewel. However, we are still > using legacy tunables. > > All our ceph infr

Re: [ceph-users] Multi-device BlueStore OSDs multiple fsck failures

2016-08-03 Thread Gregory Farnum
On Wed, Aug 3, 2016 at 3:50 PM, Somnath Roy wrote: > Probably, it is better to move to latest master and reproduce this defect. > Lot of stuff has changed between this. > This is a good test case and I doubt any of us testing by enabling fsck() on > mount/unmount. Given that the allocator keeps

Re: [ceph-users] Advice on migrating from legacy tunables to Jewel tunables.

2016-08-08 Thread Gregory Farnum
On Thu, Aug 4, 2016 at 8:57 PM, Goncalo Borges wrote: > Dear cephers... > > I am looking for some advice on migrating from legacy tunables to Jewel > tunables. > > What would be the best strategy? > > 1) A step by step approach? > - starting with the transition from bobtail to firefly (and, in

Re: [ceph-users] Advice on migrating from legacy tunables to Jewel tunables.

2016-08-08 Thread Gregory Farnum
nk we should set? > > What do you reckon? That is generally the big one, but I think you'll need advice from people who actually run clusters to see if there's anything more that's useful. :) -Greg > > Cheers > > Goncalo > > > On 08/09/2016 09:26 AM, Gre

Re: [ceph-users] how to debug pg inconsistent state - no ioerrors seen

2016-08-09 Thread Gregory Farnum
On Tue, Aug 9, 2016 at 2:00 AM, Kenneth Waegeman wrote: > Hi, > > I did a diff on the directories of all three the osds, no difference .. So I > don't know what's wrong. omap (as implied by the omap_digest complaint) is stored in the OSD leveldb, not in the data directories, so you wouldn't expec

Re: [ceph-users] Differences between default config values and docs

2013-02-24 Thread Gregory Farnum
On Sunday, February 24, 2013 at 1:48 PM, Andrey Korolyov wrote: > Hi, > > "filestore_queue_max_ops": "50", > "filestore_flusher": "true", > > Just checked that this options are set with values above instead of > config` 500 and 0 as mentioned in docs, and second option has an > observable perfo

Re: [ceph-users] Differences between default config values and docs

2013-02-24 Thread Gregory Farnum
On Sunday, February 24, 2013 at 2:42 PM, Andrey Korolyov wrote: > On Mon, Feb 25, 2013 at 2:37 AM, Gregory Farnum (mailto:g...@inktank.com)> wrote: > > On Sunday, February 24, 2013 at 1:48 PM, Andrey Korolyov wrote: > > > Hi, > > > > > > "filestore_qu

Re: [ceph-users] Differences between default config values and docs

2013-02-24 Thread Gregory Farnum
On Sunday, February 24, 2013 at 2:59 PM, Andrey Korolyov wrote: > On Mon, Feb 25, 2013 at 2:49 AM, Gregory Farnum (mailto:g...@inktank.com)> wrote: > > On Sunday, February 24, 2013 at 2:42 PM, Andrey Korolyov wrote: > > > On Mon, Feb 25, 2013 at 2:37 AM, Gregory

Re: [ceph-users] Differences between default config values and docs

2013-02-25 Thread Gregory Farnum
On Sun, Feb 24, 2013 at 3:10 PM, Andrey Korolyov wrote: > On Mon, Feb 25, 2013 at 3:07 AM, Gregory Farnum wrote: >> Try >> ceph --keyfile admin osd tell \* injectargs "--filestore-flusher=0" >> (add the equals sign) and let me know if that works. Looks like may

Re: [ceph-users] Differences between default config values and docs

2013-02-25 Thread Gregory Farnum
On Mon, Feb 25, 2013 at 11:14 AM, Gregory Farnum wrote: > On Sun, Feb 24, 2013 at 3:10 PM, Andrey Korolyov wrote: >> On Mon, Feb 25, 2013 at 3:07 AM, Gregory Farnum wrote: >>> Try >>> ceph --keyfile admin osd tell \* injectargs "--filestore-flusher=0" >

Re: [ceph-users] mds crash

2013-02-25 Thread Gregory Farnum
On Mon, Feb 25, 2013 at 8:44 AM, Sage Weil wrote: > On Mon, 25 Feb 2013, Steffen Thorhauer wrote: >> Hi, >> I've found out, what I make wrong: stop the cluster and forget a client, >> which as mounting the cephfs. I simply forget the client. >> With a >> ceph mds newfs 0 1 --yes-i-really-mean-it

Re: [ceph-users] Resizing pool of pg (.rgw*)

2013-02-28 Thread Gregory Farnum
On Wed, Feb 27, 2013 at 3:20 AM, Yann ROBIN wrote: > Hi, > > What could go wrong with this command ? If something goes wrong, is there any > chance of recovery ? > As this is a production cluster, I don't want to test it if there is a chance > of losing my cluster an data. It could go horribly

Re: [ceph-users] 0.56.3 OSDs wrongly marked down and cluster unresponsiveness

2013-02-28 Thread Gregory Farnum
I haven't checked the logs at all, but my first question is what the memory utilization on that node looked like during that time, and if there are any dmesg warnings. (Other than that I should leave this to probably Alex and Sam.) -Greg On Thu, Feb 28, 2013 at 1:44 PM, Nick Bartos wrote: > When

Re: [ceph-users] Placement Groups

2013-03-01 Thread Gregory Farnum
On Fri, Mar 1, 2013 at 8:17 AM, Scott Kinder wrote: > In my ceph.conf file, I set the options under the [osd] section: > > osd pool default pg num = 133 > osd pool default pgp num = 133 You'll want to place those in the global section rather than the OSD section — the pools are generated on the m

Re: [ceph-users] Public network

2013-03-01 Thread Gregory Farnum
This doesn't cover reads from the same objects though, and in fact Ceph deliberately sends all object reads to the primary copy. There are various options (which are variously implemented) to turn this off but turning it into a proper striped reads system or something would take some work. -Greg O

Re: [ceph-users] Concurrent access to Ceph filesystems

2013-03-01 Thread Gregory Farnum
On Fri, Mar 1, 2013 at 1:53 PM, Karsten Becker wrote: > Hi, > > I'm new to Ceph. I currently find no answer in the official docs for the > following question. > > Can Ceph filesystems be used concurrently by clients, both when > accessing via RBD and CephFS? Concurrently means in terms of multiple

Re: [ceph-users] Concurrent access to Ceph filesystems

2013-03-01 Thread Gregory Farnum
This doesn't sound quite right but I'm not sure if the problem is a terminology disconnect or a concept one. Let's go through them On Fri, Mar 1, 2013 at 3:08 PM, McNamara, Bradley wrote: > I'm new, too, and I guess I just need a little clarification on Greg's > statement. The RBD filesystem >

Re: [ceph-users] Changing CRUSH rule on a running cluster

2013-03-04 Thread Gregory Farnum
On Mon, Mar 4, 2013 at 12:19 AM, Olivier Bonvalet wrote: > Hello, > > I have a running cluster, which use the (previous) default CRUSH rule, > with : > step take default > step choose firstn 0 type osd > step emit > > Since I have multiple OSD in the same host, I need to ch

Re: [ceph-users] Crashed MDS not starting

2013-03-06 Thread Gregory Farnum
On Wednesday, March 6, 2013 at 2:43 AM, Félix Ortega Hortigüela wrote: > Hi > I'm running ceph v56.3 over debian-wheezy, with the ceph.com > (http://ceph.com) debs. > My setup is three servers with 6 disk each. I have 5 disks on each server > dedicated to osd's and the disk left is dedicated to t

Re: [ceph-users] If one node lost connect to replication network?

2013-03-11 Thread Gregory Farnum
On Monday, March 11, 2013, Igor Laskovy wrote: > Hi there! > > I have Ceph FS cluster version 0.56.3. This is 3 nodes with XFS on disks > and with minimum options in ceph.conf in my lab and I do some crush > testing. > One of the of several tests is lost connect to replication network only. > What

Re: [ceph-users] Uneven data placement

2013-03-17 Thread Gregory Farnum
On Sunday, March 17, 2013 at 4:46 AM, Andrey Korolyov wrote: > Hi, > > from osd tree: > > -16 4.95 host 10.5.0.52 > 32 1.9 osd.32 up 2 > 33 1.05 osd.33 up 1 > 34 1 osd.34 up 1 > 35 1 osd.35 up 1 > > df -h: > /dev/sdd3 3.7T 595G 3.1T 16% /var/lib/ceph/osd/32 > /dev/sde3 3.7T 332G 3.4T 9% /var/

Re: [ceph-users] Uneven data placement

2013-03-17 Thread Gregory Farnum
On Sunday, March 17, 2013 at 9:09 AM, Andrey Korolyov wrote: > On Sun, Mar 17, 2013 at 7:56 PM, Gregory Farnum (mailto:g...@inktank.com)> wrote: > > On Sunday, March 17, 2013 at 4:46 AM, Andrey Korolyov wrote: > > > Hi, > > > > > > from osd tree: > &g

Re: [ceph-users] how to enable MDS service in a running Ceph cluster

2013-03-17 Thread Gregory Farnum
On Friday, March 15, 2013 at 2:02 AM, Li, Chen wrote: > I need to create the directory “/var/lib/ceph/mds/mds.$id ”by hand, right ? > > I start the service as you said, and it is succeed. > But, no “mds.$id” directory exist. > Will this affect it working? > > And, what will be installed in the

Re: [ceph-users] Planning for many small files

2013-03-17 Thread Gregory Farnum
RADOS doesn't store a list of objects. The RADOS Gateway uses a separate data format on top of objects stored in RADOS, and it keeps a per-user list of buckets and a per-bucket index of objects as "omap" objects in the OSDs (which ultimately end up in a leveldb store). A bucket index is currentl

Re: [ceph-users] Uneven data placement

2013-03-17 Thread Gregory Farnum
On Sunday, March 17, 2013 at 9:25 AM, Andrey Korolyov wrote: > On Sun, Mar 17, 2013 at 8:14 PM, Gregory Farnum (mailto:g...@inktank.com)> wrote: > > On Sunday, March 17, 2013 at 9:09 AM, Andrey Korolyov wrote: > > > On Sun, Mar 17, 2013 at 7:56 PM, Gregory Farnum > &g

Re: [ceph-users] using ssds with ceph

2013-03-17 Thread Gregory Farnum
On Sunday, March 17, 2013 at 4:03 PM, Mark Nelson wrote: > On 03/17/2013 05:40 PM, Matthieu Patou wrote: > > Hello all, > > > > Our dev environment are quite I/O intensive but didn't require much > > space (~20G per dev environment), for the moment our dev machines are > > served by VMWare and th

Re: [ceph-users] Planning for many small files

2013-03-17 Thread Gregory Farnum
On Sunday, March 17, 2013 at 5:49 PM, Yehuda Sadeh wrote: > Not at the moment. We had some discussions about "blind" buckets, it's > definitely on our mind, but we're not there yet. > > Yehuda > > On Sun, Mar 17, 2013 at 3:09 PM, Rustam Aliyev (mailto:rustam.li...@code.az)> wrote: > > Thanks for

Re: [ceph-users] Crush map example

2013-03-20 Thread Gregory Farnum
On Wed, Mar 20, 2013 at 5:06 PM, Darryl Bond wrote: > I have a cluster of 3 hosts each with 2 SSD and 4 Spinning disks. > I used the example in th ecrush map doco to create a crush map to place > the primary on the SSD and replica on spinning disk. > > If I use the example, I end up with objects r

Re: [ceph-users] Journal size

2013-03-27 Thread Gregory Farnum
This should actually be "down", not "out". Out will trigger a data migration. :) -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Wed, Mar 27, 2013 at 8:36 AM, John Nielsen wrote: > Sébastien Han actually covers this on his blog: > > http://www.sebastien-han.fr/blog/2012/08

Re: [ceph-users] Cluster Map Problems

2013-03-28 Thread Gregory Farnum
This is the perfectly normal distinction between "down" and "out". The OSD has been marked down but there's a timeout period (default: 5 minutes) before it's marked "out" and the data gets reshuffled (to avoid starting replication on a simple reboot, for instance). -Greg Software Engineer #42 @ htt

Re: [ceph-users] Cluster Map Problems

2013-03-28 Thread Gregory Farnum
168.195.33:6789/0,c=192.168.195.35:6789/0}, > election epoch 6, quorum 0,1,2 a,b,c >osdmap e28: 24 osds: 23 up, 24 in > pgmap v454: 4800 pgs: 4422 active+clean, 378 active+degraded; 1800 > MB data, 3780 MB used, 174 TB / 174 TB avail; 39/904 degraded (4.314%) >mdsmap e1

Re: [ceph-users] Cluster Map Problems

2013-03-28 Thread Gregory Farnum
default > step chooseleaf firstn 0 type rack > step emit > } > rule metadata { > ruleset 1 > type replicated > min_size 1 > max_size 10 > step take default > step chooseleaf firstn 0 type rack > step

Re: [ceph-users] Cluster Map Problems

2013-03-28 Thread Gregory Farnum
/dev/sdd >>> [osd.14] >>> host = store4 >>> osd journal = /dev/sdi1 >>> btrfs devs = /dev/sde >>> [osd.15] >>> host = store4 >>> osd journal = /dev/sdj1 >>> btrfs devs = /dev/sdf >&

Re: [ceph-users] Newbie questions re: ceph setup

2013-04-01 Thread Gregory Farnum
On Mon, Apr 1, 2013 at 2:16 PM, Sam Lang wrote: > On Mon, Apr 1, 2013 at 5:59 AM, Papaspyrou, Alexander > wrote: >> Folks, >> >> we are trying to setup a ceph cluster with about 40 or so OSDs on our >> hosting provider's infrastructure. Our rollout works with Opscode Chef, and >> I'm driving my p

Re: [ceph-users] Slow RBD copy

2013-04-02 Thread Gregory Farnum
On Sat, Mar 30, 2013 at 3:46 AM, Wido den Hollander wrote: > On 03/29/2013 01:42 AM, Steve Carter wrote: >> >> I create an empty 150G volume them copy it to a second pool: >> >> # rbd -p pool0 create --size 153750 steve150 >> >> # /usr/bin/time rbd cp pool0/steve150 pool1/steve150 >> Image copy: 1

Re: [ceph-users] Newbie questions re: ceph setup

2013-04-03 Thread Gregory Farnum
On Wed, Apr 3, 2013 at 9:45 AM, John Nielsen wrote: > On Apr 1, 2013, at 3:33 PM, Gregory Farnum wrote: > >>> On Mon, Apr 1, 2013 at 2:16 PM, Sam Lang wrote: >>>> On Mon, Apr 1, 2013 at 5:59 AM, Papaspyrou, Alexander >>>> wrote: >>>> 1. S

Re: [ceph-users] Ceph 0.60 showing data in cluster that does not exist

2013-04-03 Thread Gregory Farnum
And if you put a big file in CephFS and then deleted it, the data will be deleted from the RADOS cluster asynchronously in the background (by the MDS), so it can take a while to actually get removed. :) If this wasn't the behavior then a file delete would require you to wait for each of those (10GB

Re: [ceph-users] Ceph mon quorum

2013-04-05 Thread Gregory Farnum
On Fri, Apr 5, 2013 at 10:28 AM, Dimitri Maziuk wrote: > On 04/05/2013 10:12 AM, Wido den Hollander wrote: > >> Think about it this way. You have two racks and the network connection >> between them fails. If both racks keep operating because they can still >> reach that single monitor in their ra

Re: [ceph-users] Performance problems

2013-04-08 Thread Gregory Farnum
More details, please. You ran the same test twice and performance went up from 17.5MB/s to 394MB/s? How many drives in each node, and of what kind? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Mon, Apr 8, 2013 at 12:38 PM, Ziemowit Pierzycki wrote: > Hi, > > I have a 3 n

Re: [ceph-users] Backup of cephfs metadata

2013-04-09 Thread Gregory Farnum
If you were to do that you'd be going backwards in time with your metadata, so — not really. CephFS is not generally production-ready at this time, but we welcome bug reports! -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Mon, Apr 8, 2013 at 12:52 PM, Maik Kulbe wrote: >

Re: [ceph-users] determine the size

2013-04-10 Thread Gregory Farnum
On Wednesday, April 10, 2013 at 2:53 AM, Waed Bataineh wrote: > Hello, > > I have several question i'll be appreciated if i got answers for them: > > 1. does the osd have a fixed size or it compatible with the machine > i'm working with. You can weight OSDs to account for different capacities or

Re: [ceph-users] ceph segfault on all osd

2013-04-10 Thread Gregory Farnum
Sounds like they aren't handling the transition very well when trying to calculate old OSDs which might have held the PG. Are you trying to salvage the data that was in it, or can you throw it away? Can you post the backtrace they're producing? -Greg Software Engineer #42 @ http://inktank.com |

Re: [ceph-users] ceph segfault on all osd

2013-04-10 Thread Gregory Farnum
size 2 > > that all. > > > > > > > On 04/10/2013 05:24 PM, Gregory Farnum wrote: > > Sounds like they aren't handling the transition very well when trying to > > calculate old OSDs which might have held the PG. Are you trying to salvage

Re: [ceph-users] ceph segfault on all osd

2013-04-10 Thread Gregory Farnum
talij Poljatchek wrote: > the log files. > > thank you ! :) > > On 04/10/2013 06:06 PM, Gregory Farnum wrote: > > [Re-adding the list.] > > > > When the OSDs crash they will print out to their log a short description of > > what happened, with a bunch of

Re: [ceph-users] Backup of cephfs metadata

2013-04-10 Thread Gregory Farnum
Well, if you've made changes to your data which impacted the metadata, and then you restore to a backup of the metadata pool, but not the data, then what's there isn't what CephFS thinks is there. Which would be confusing for all the same reasons that it is in a local filesystem. You could construc

Re: [ceph-users] Backup of cephfs metadata

2013-04-10 Thread Gregory Farnum
We've discussed the order of work (you can see my recent Ceph blog post on the subject; though it's subject to revision) but haven't committed to any dates at this time. Sorry. :( -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Wed, Apr 10, 2013 at 12:43 PM, Maik Kulbe wrot

Re: [ceph-users] pool-info error in version 0.60

2013-04-11 Thread Gregory Farnum
It's more or less a Ceph bug; the patch fixing this is in the 3.9-rc's (although it should backport trivially if you're willing to build a kernel: 92a49fb0f79f3300e6e50ddf56238e70678e4202). You can look at http://tracker.ceph.com/issues/3793 if you want details. -Greg Software Engineer #42 @ http:/

Re: [ceph-users] MDS crashed (ver 0.56.2)

2013-04-11 Thread Gregory Farnum
That's certainly not great. Have you lost any data or removed anything from the cluster? It looks like perhaps your MDS log lost an object, and maybe got one shortened as well. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Mon, Apr 8, 2013 at 11:55 PM, x yasha wrote: > I'

Re: [ceph-users] Request for Ceph peering state model graph

2013-04-12 Thread Gregory Farnum
Sam was complaining about this a few days ago too; I'm not sure if he ever found a solution. If you've got the patience you can actually build that image yourself; the source is in our git repo. If you don't, I'm sure our intrepid doc/website people can get it fixed up fairly soon. :) -Greg Softwar

Re: [ceph-users] ceph -w question

2013-04-15 Thread Gregory Farnum
"Incomplete" means that there are fewer than the minimum copies of the placement group (by default, half of the requested size, rounded up). In general rebooting one node shouldn't do that unless you've changed your minimum size on the pool, and it does mean that data in those PGs is unaccessible.

Re: [ceph-users] Scrub shutdown the OSD process

2013-04-15 Thread Gregory Farnum
On Mon, Apr 15, 2013 at 2:42 AM, Olivier Bonvalet wrote: > Hi, > > I have an OSD process which is regulary shutdown by scrub, if I well > understand that trace : > > 0> 2013-04-15 09:29:53.708141 7f5a8e3cc700 -1 *** Caught signal > (Aborted) ** > in thread 7f5a8e3cc700 > > ceph version 0.5

Re: [ceph-users] Scrub shutdown the OSD process

2013-04-15 Thread Gregory Farnum
On Mon, Apr 15, 2013 at 10:19 AM, Olivier Bonvalet wrote: > Le lundi 15 avril 2013 à 10:16 -0700, Gregory Farnum a écrit : >> Are you saying you saw this problem more than once, and so you >> completely wiped the OSD in question, then brought it back into the >> cluster, and

Re: [ceph-users] cephfs bandwidth issue

2013-04-16 Thread Gregory Farnum
On Tue, Apr 16, 2013 at 8:47 AM, Elso Andras wrote: > Hi, > > We are serving mp4 videos from cephfs, with lots of random read. > > Here is a graph: http://elbandi.net/ceph/ceph_bandwidth.png > The issue is: there are +50% incoming bandwidth (osd -> cephfs mount, blue > line) compared outgoing band

Re: [ceph-users] Format 2 Image support in the RBD driver

2013-04-18 Thread Gregory Farnum
I believe Alex just merged format 2 reading into our testing branch, and is working on writes now. -Greg On Thursday, April 18, 2013, Whelan, Ryan wrote: > Does this mean its in linux-next? (released in 3.10?) > > - Original Message - > From: "Olivier B." > > To: "Ryan Whelan" > > Cc: cep

Re: [ceph-users] spontaneous pg inconstancies in the rgw.gc pool

2013-04-18 Thread Gregory Farnum
What version was this on? -Greg On Thursday, April 18, 2013, Dan van der Ster wrote: > Sorry for the noise.. we now have a better idea what happened here. > > For those that might care, basically we had one client looping while > trying to list the / bucket with an incorrect key. rgw was handling

Re: [ceph-users] has anyone successfully installed ceph with the crowbar

2013-04-18 Thread Gregory Farnum
The barclamps were written against the crowbar "Betty" release, OpenStack Essex (which is the last one supported by Crowbar), and Ceph "argonaut". JJ has updated them to use "Bobtail", but I don't think anybody's run them against newer versions of Openstack. :( You should be able to find built vers

Re: [ceph-users] has anyone successfully installed ceph with the crowbar

2013-04-18 Thread Gregory Farnum
the volumes pool it fails. > this is due to some permission failures probably because i try to follow a > manual that is created for bobtail > http://ceph.com/docs/master/rbd/rbd-openstack/ > > > > > On Thu, 2013-04-18 at 08:48 -0700, Gregory Farnum wrote: > > The

Re: [ceph-users] No rolling updates from v0.56 to v0.60+?

2013-04-18 Thread Gregory Farnum
On Wed, Apr 17, 2013 at 7:40 AM, Guido Winkelmann wrote: > Hi, > > I just tried upgrading parts of our experimental ceph cluster from 0.56.1 to > 0.60, and it looks like the new mon-daemon from 0.60 cannot talk to those from > 0.56.1 at all. > > Long story short, we had to move some hardware aroun

Re: [ceph-users] Bobtail & Precise

2013-04-18 Thread Gregory Farnum
Seeing this go by again it's simple enough to provide a quick answer/hint — by setting the tunables it's of course getting a better distribution of data, but the reason they're optional to begin with is that older clients won't support them. In this case, the kernel client being run; so it returns

Re: [ceph-users] Bobtail & Precise

2013-04-18 Thread Gregory Farnum
ell. > > > On Thu, Apr 18, 2013 at 1:56 PM, Gregory Farnum wrote: > >> Seeing this go by again it's simple enough to provide a quick >> answer/hint — by setting the tunables it's of course getting a better >> distribution of data, but the reason they're

Re: [ceph-users] Monitor Access Denied message to itself?

2013-04-18 Thread Gregory Farnum
Hey guys, I finally had enough time to coordinate with a few other people and figure out what's going on with the ceph-create-keys access denied messages and create a ticket: http://tracker.ceph.com/issues/4752. (I believe your monitor crash is something else, Matthew; if that hasn't been dealt wit

Re: [ceph-users] Monitor Access Denied message to itself?

2013-04-18 Thread Gregory Farnum
On Thu, Apr 18, 2013 at 2:46 PM, Joao Eduardo Luis wrote: > On 04/18/2013 10:36 PM, Gregory Farnum wrote: >> >> (I believe your monitor crash is something else, Matthew; if that >> hasn't been dealt with yet. Unfortunately all that log has is >> messages, so it pr

Re: [ceph-users] Monitor Access Denied message to itself?

2013-04-18 Thread Gregory Farnum
There's a little bit of python called ceph-create-keys, which is invoked by the upstart scripts. You can kill the running processes, and edit them out of the scripts, without direct harm. (Their purpose is to create some standard keys which the newer deployment tools rely on to do things like creat

Re: [ceph-users] Monitor Access Denied message to itself?

2013-04-19 Thread Gregory Farnum
On Thu, Apr 18, 2013 at 7:59 PM, Mike Dawson wrote: > Greg, > > Looks like Sage has a fix for this problem. In case it matters, I have seen > a few cases that conflict with your notes in this thread and the bug report. > > I have seen the bug exclusively on new Ceph installs (without upgrading fro

Re: [ceph-users] Monitor Access Denied message to itself?

2013-04-19 Thread Gregory Farnum
On Fri, Apr 19, 2013 at 9:11 AM, Mike Dawson wrote: > > > On 4/19/2013 11:43 AM, Gregory Farnum wrote: >> >> On Thu, Apr 18, 2013 at 7:59 PM, Mike Dawson >> wrote: >>> >>> Greg, >>> >>> Looks like Sage has a fix for this problem. In

Re: [ceph-users] Stalled requests; help?

2013-04-19 Thread Gregory Farnum
On Fri, Apr 19, 2013 at 3:12 PM, Steven Presser wrote: > Hey all, > I've got a ceph cluster set up (0.56.4) on a custom centos image (base > centos 6, plus kernel 3.6.9) running as a Xen dom0. I'm seeing a lot of > messages like the ones at the bottom of this message. I'm entirely willing >

Re: [ceph-users] Stalled requests; help?

2013-04-19 Thread Gregory Farnum
tty early in the setup. (Heck, i just got it switched to > gigabit ethernet...) > > Thanks, > Steve > > > On 04/19/2013 06:31 PM, Gregory Farnum wrote: >> >> On Fri, Apr 19, 2013 at 3:12 PM, Steven Presser >> wrote: >>> >>> Hey all, >>>

Re: [ceph-users] clean shutdown and failover of osd

2013-04-20 Thread Gregory Farnum
[ This is a good query for ceph-users. ] On Sat, Apr 20, 2013 at 10:15 PM, James Harper wrote: > I'm doing some testing with ceph trying to figure out why my performance is > so bad, and have noticed that there doesn't seem to be a way to cleanly stop > an osd, or at least under debian /etc/ini

Re: [ceph-users] Cephfs unaccessible

2013-04-20 Thread Gregory Farnum
It's not entirely clear from your description and the output you've given us, but it looks like maybe you've managed to bring up all your OSDs correctly at this point? Or are they just not reporting down because you set the "no down" flag... In any case, CephFS isn't going to come up while the und

<    9   10   11   12   13   14   15   16   17   18   >