Re: [ceph-users] inotify, etc?

2015-10-23 Thread Gregory Farnum
On Fri, Oct 23, 2015 at 10:14 AM, Edward Ned Harvey (ceph) wrote: > Trying to figure out if ceph supports inotify, or some form of notification, > I see this issue from 4 years ago: > > http://tracker.ceph.com/issues/1296 > > > > And the corresponding discussion thread > > http://comments.gmane.or

Re: [ceph-users] Our 0.94.2 OSD are not restarting : osd/PG.cc: 2856: FAILED assert(values.size() == 1)

2015-10-27 Thread Gregory Farnum
You might see if http://tracker.ceph.com/issues/13060 could apply to your cluster. If so upgrading to .94.4 should fix it. *Don't* reset your OSD journal. That is never the answer and is basically the same as trashing the OSD in question. -Greg On Tue, Oct 27, 2015 at 9:59 AM, Laurent GUERBY wro

Re: [ceph-users] PGs stuck in active+clean+replay

2015-10-27 Thread Gregory Farnum
On Thu, Oct 22, 2015 at 3:58 PM, Andras Pataki wrote: > Hi ceph users, > > We’ve upgraded to 0.94.4 (all ceph daemons got restarted) – and are in the > middle of doing some rebalancing due to crush changes (removing some disks). > During the rebalance, I see that some placement groups get stuck in

Re: [ceph-users] PGs stuck in active+clean+replay

2015-10-27 Thread Gregory Farnum
On Tue, Oct 27, 2015 at 11:03 AM, Gregory Farnum wrote: > On Thu, Oct 22, 2015 at 3:58 PM, Andras Pataki > wrote: >> Hi ceph users, >> >> We’ve upgraded to 0.94.4 (all ceph daemons got restarted) – and are in the >> middle of doing some rebalancing due to crus

Re: [ceph-users] PGs stuck in active+clean+replay

2015-10-27 Thread Gregory Farnum
On Tue, Oct 27, 2015 at 11:22 AM, Andras Pataki wrote: > Hi Greg, > > No, unfortunately I haven¹t found any resolution to it. We are using > cephfs, the whole installation is on 0.94.4. What I did notice is that > performance is extremely poor when backfilling is happening. I wonder if > timeou

Re: [ceph-users] values of "ceph daemon osd.x perf dump objecters " are zero

2015-10-28 Thread Gregory Farnum
[ Removed ceph-devel ] On Wednesday, October 28, 2015, Libin Wu wrote: > Hi, all > > As my understand, command "ceph daemon osd.x perf dump objecters" should > output the perf data of osdc(librados). But when i use this command, > why all those values are zero expcept map_epoch and map_inc. Foll

Re: [ceph-users] CephFS and page cache

2015-10-28 Thread Gregory Farnum
On Wed, Oct 28, 2015 at 8:38 PM, Yan, Zheng wrote: > On Thu, Oct 29, 2015 at 1:10 AM, Burkhard Linke >> I tried to dig into the ceph-fuse code, but I was unable to find the >> fragment that is responsible for flushing the data from the page cache. >> > > fuse kernel code invalidates page cache on

Re: [ceph-users] Benchmark individual OSD's

2015-10-29 Thread Gregory Farnum
You can also extend that command line to specify specific block and total sizes. Check the help text. :) -Greg On Thursday, October 29, 2015, Lindsay Mathieson < lindsay.mathie...@gmail.com> wrote: > > On 29 October 2015 at 19:24, Burkhard Linke < > burkhard.li...@computational.bio.uni-giessen.de

Re: [ceph-users] SHA1 wrt hammer release and tag v0.94.3

2015-10-30 Thread Gregory Farnum
On Fri, Oct 30, 2015 at 6:20 PM, Artie Ziff wrote: > Hello, > > In the RELEASE INFORMATION section of the hammer v0.94.3 issue tracker [1] > the git commit SHA1 is: b2503b0e15c0b13f480f0835060479717b9cf935 > > On the github page for Ceph Release v0.94.3 [2], when I click on the > "95cefea" link [3

Re: [ceph-users] data size less than 4 mb

2015-10-31 Thread Gregory Farnum
On Friday, October 30, 2015, mad Engineer wrote: > i am learning ceph,block storage and read that each object size is 4 Mb.I > am not clear about the concepts of object storage still what will happen if > the actual size of data written to block is less than 4 Mb lets say 1 > Mb.Will it still cre

Re: [ceph-users] Changing CRUSH map ids

2015-11-02 Thread Gregory Farnum
Regardless of what the crush tool does, I wouldn't muck around with the IDs of the OSDs. The rest of Celh will probably not handle it well if the crush IDs don't match the OSD numbers. -Greg On Monday, November 2, 2015, Loris Cuoghi wrote: > Le 02/11/2015 12:47, Wido den Hollander a écrit : > >>

Re: [ceph-users] Changing CRUSH map ids

2015-11-02 Thread Gregory Farnum
, and testing the two > maps with : > > crushtool -i crush.map --test --show-statistics --rule 0 --num-rep 3 --min-x > 1 --max-x $N --show-mappings > > (with $N varying from as little as 32 to "big numbers"TM) shows that nearly > the 50% of the mappings changed, in a 10 host

Re: [ceph-users] rados bench leaves objects in tiered pool

2015-11-03 Thread Gregory Farnum
When you have a caching pool in writeback mode, updates to objects (including deletes) are handled by writeback rather than writethrough. Since there's no other activity against these pools, there is nothing prompting the cache pool to flush updates out to the backing pool, so the backing pool hasn

Re: [ceph-users] rados bench leaves objects in tiered pool

2015-11-03 Thread Gregory Farnum
11, Robert LeBlanc > написал(а): > > > > -----BEGIN PGP SIGNED MESSAGE- > > Hash: SHA256 > > > > Try: > > > > rados -p {cachepool} cache-flush-evict-all > > > > and see if the objects clean up. > > - >

Re: [ceph-users] Increased pg_num and pgp_num

2015-11-04 Thread Gregory Farnum
It shouldn't be -- if you changed pg_num then a bunch of PGs will need to move and will report in this state. We can check more thoroughly if you provide the full "Ceph -s" output. (Stuff to check for: that all PGs are active, none are degraded, etc) -Greg On Wednesday, November 4, 2015, Erming Pe

Re: [ceph-users] Understanding the number of TCP connections between clients and OSDs

2015-11-04 Thread Gregory Farnum
On Wed, Nov 4, 2015 at 12:27 PM, Rick Balsano wrote: > Just following up since this thread went silent after a few comments showing > similar concerns, but no explanation of the behavior. Can anyone point to > some code or documentation which explains how to estimate the expected > number of TCP c

Re: [ceph-users] Soft removal of RBD images

2015-11-06 Thread Gregory Farnum
On Fri, Nov 6, 2015 at 2:03 AM, Wido den Hollander wrote: > Hi, > > Since Ceph Hammer we can protect pools from being removed from the > cluster, but we can't protect against this: > > $ rbd ls|xargs -n 1 rbd rm > > That would remove all not opened RBD images from the cluster. > > This requires di

Re: [ceph-users] osd fails to start, rbd hangs

2015-11-06 Thread Gregory Farnum
http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshooting-pg/ :) On Friday, November 6, 2015, Philipp Schwaha wrote: > Hi, > > I have an issue with my (small) ceph cluster after an osd failed. > ceph -s reports the following: > cluster 2752438a-a33e-4df4-b9ec-beae32d00aad >

Re: [ceph-users] Erasure coded pools and 'feature set mismatch' issue

2015-11-08 Thread Gregory Farnum
With that release it shouldn't be the EC pool causing trouble; it's the CRUSH tunables also mentioned in that thread. Instructions should be available in the docs for using older tunable that are compatible with kernel 3.13. -Greg On Saturday, November 7, 2015, Bogdan SOLGA wrote: > Hello, every

Re: [ceph-users] cephfs: Client hp-s3-r4-compute failing to respondtocapabilityrelease

2015-11-09 Thread Gregory Farnum
On Mon, Nov 9, 2015 at 6:57 AM, Burkhard Linke wrote: > Hi, > > On 11/09/2015 02:07 PM, Burkhard Linke wrote: >> >> Hi, > > *snipsnap* > >> >> >> Cluster is running Hammer 0.94.5 on top of Ubuntu 14.04. Clients use >> ceph-fuse with patches for improved page cache handling, but the problem >> also

Re: [ceph-users] Seeing which Ceph version OSD/MON data is

2015-11-09 Thread Gregory Farnum
The daemons print this in their debug logs on every boot. (There might be a minimum debug level required, but I think it's at 0!) -Greg On Mon, Nov 9, 2015 at 7:23 AM, Wido den Hollander wrote: > Hi, > > Recently I got my hands on a Ceph cluster which was pretty damaged due > to a human error. >

Re: [ceph-users] crush rule with two parts

2015-11-09 Thread Gregory Farnum
On Mon, Nov 9, 2015 at 9:42 AM, Deneau, Tom wrote: > I don't have much experience with crush rules but wanted one that does the > following: > > On a 3-node cluster, I wanted a rule where I could have an erasure-coded pool > of k=3,m=2 > and where the first 3 chunks (the read chunks) are all on

Re: [ceph-users] cephfs: Client hp-s3-r4-compute failing torespondtocapabilityrelease

2015-11-10 Thread Gregory Farnum
Can you dump the metadata ops in flight on each ceph-fuse when it hangs? ceph daemon mds_requests -Greg On Mon, Nov 9, 2015 at 8:06 AM, Burkhard Linke wrote: > Hi, > > On 11/09/2015 04:03 PM, Gregory Farnum wrote: >> >> On Mon, Nov 9, 2015 at 6:57 AM, Burkhard Linke

Re: [ceph-users] Permanent MDS restarting under load

2015-11-10 Thread Gregory Farnum
On Tue, Nov 10, 2015 at 6:32 AM, Oleksandr Natalenko wrote: > Hello. > > We have CephFS deployed over Ceph cluster (0.94.5). > > We experience constant MDS restarting under high IOPS workload (e.g. > rsyncing lots of small mailboxes from another storage to CephFS using > ceph-fuse client). First,

Re: [ceph-users] Using straw2 crush also with Hammer

2015-11-11 Thread Gregory Farnum
On Wednesday, November 11, 2015, Wido den Hollander wrote: > On 11/10/2015 09:49 PM, Vickey Singh wrote: > > On Mon, Nov 9, 2015 at 8:16 PM, Wido den Hollander > wrote: > > > >> On 11/09/2015 05:27 PM, Vickey Singh wrote: > >>> Hello Ceph Geeks > >>> > >>> Need your comments with my understandin

Re: [ceph-users] Ceph file system is not freeing space

2015-11-11 Thread Gregory Farnum
On Wed, Nov 11, 2015 at 2:28 PM, Eric Eastman wrote: > On Wed, Nov 11, 2015 at 11:09 AM, John Spray wrote: >> On Wed, Nov 11, 2015 at 5:39 PM, Eric Eastman >> wrote: >>> I am trying to figure out why my Ceph file system is not freeing >>> space. Using Ceph 9.1.0 I created a file system with sna

Re: [ceph-users] Ceph object mining

2015-11-13 Thread Gregory Farnum
I think I saw somebody working on a RADOS interface to Apache Hadoop once, maybe search for that? Your other option is to try and make use of object classes directly, but that's a bit orimitive to build full map-reduce on top of without a lot of effort. -Greg On Friday, November 13, 2015, min fang

Re: [ceph-users] all pgs of erasure coded pool stuck stale

2015-11-13 Thread Gregory Farnum
Somebody else will need to do the diagnosis, but it'll help them if you can get logs with "debug ms = 1", "debug osd = 20" in the log. Based on the required features update in the crush map, it looks like maybe you've upgraded some of your OSDs — is that a thing happening right now? Perhaps you up

Re: [ceph-users] pg stuck in remapped+peering for a long time

2015-11-14 Thread Gregory Farnum
What's the full output of "Ceph -s"? Are your new crush rules actually satisfiable? Is your cluster filling up? -Greg On Saturday, November 14, 2015, Peter Theobald wrote: > Hi list, > > I have a 3 node ceph cluster with a total of 9 ods (2,3 and 4 with > different size drives). I changed the la

Re: [ceph-users] rados_aio_cancel

2015-11-17 Thread Gregory Farnum
On Monday, November 16, 2015, min fang wrote: > Is this function used in detach rx buffer, and complete IO back to the > caller? From the code, I think this function will not interact with OSD or > MON side, which means, we just cancel IO from client side. Am I right? > > Thanks. > Right. If th

Re: [ceph-users] ceph infernalis pg creating forever

2015-11-20 Thread Gregory Farnum
This usually means your crush mapping for the pool in question is unsatisfiable. Check what the rule is doing. -Greg On Friday, November 20, 2015, German Anders wrote: > Hi all, I've finished the install of a new ceph cluster with infernalis > 9.2.0 release. But I'm getting the following error m

Re: [ceph-users] Fixing inconsistency

2015-11-23 Thread Gregory Farnum
On Wed, Nov 18, 2015 at 4:34 AM, Межов Игорь Александрович wrote: > Hi! > > As for my previous message, digging mailing list gave me only one method to > fix > inconsistency - truncate object files in a filesystem to a size, that they > have > in ceph metadata: > > http://www.spinics.net/lists/c

Re: [ceph-users] Objects per PG skew warning

2015-11-23 Thread Gregory Farnum
On Thu, Nov 19, 2015 at 8:56 PM, Richard Gray wrote: > Hi, > > Running 'health detail' on our Ceph cluster this morning, I notice a warning > about one of the pools having significantly more objects per placement group > than the cluster average. > > ceph> health detail > HEALTH_WARN pool cas_back

Re: [ceph-users] Ceph 0.94.5 with accelio

2015-11-23 Thread Gregory Farnum
On Mon, Nov 23, 2015 at 10:05 AM, German Anders wrote: > Hi all, > > I want to know if there's any improvement or update regarding ceph 0.94.5 > with accelio, I've an already configured cluster (with no data on it) and I > would like to know if there's a way to 'modify' the cluster in order to use

Re: [ceph-users] op sequence

2015-11-23 Thread Gregory Farnum
On Mon, Nov 23, 2015 at 8:44 AM, louis wrote: > Hi, if I submit read or write io in a sequence from a ceph client, will > these sequence will be kept in osds side? Thanks Any writes from the same client, to the same object, will be ordered with respect to one another. But there are no other guara

Re: [ceph-users] ceph-mon cpu 100%

2015-11-23 Thread Gregory Farnum
Yep. I think you can inject it into the running mons without restarting as well (injectargs). -Greg On Mon, Nov 23, 2015 at 11:46 AM, Yujian Peng wrote: > The mons in my production cluster have a very high cpu usage 100%. > I think it may be caused by the leveldb compression. > How yo disable lev

Re: [ceph-users] High load during recovery (after disk placement)

2015-11-23 Thread Gregory Farnum
On Fri, Nov 20, 2015 at 11:33 AM, Simon Engelsman wrote: > Hi, > > We've experienced a very weird problem last week with our Ceph > cluster. We would like to ask your opinion(s) and advice > > Our dedicated Ceph OSD nodes run with: > > Total platform > - IO Average: 2500 wrps, ~ 600 rps > - Replic

Re: [ceph-users] CACHEMODE_READFORWARD doesn't try proxy write?

2015-11-23 Thread Gregory Farnum
Yeah, the write proxying is pretty new and the fact that it's missing from an oddball like READFORWARD isn't surprising. (Not good, exactly, but not surprising.) What are you doing with this caching mode? On Thu, Nov 19, 2015 at 10:34 AM, Nick Fisk wrote: > Don’t know why that URL got changed, i

Re: [ceph-users] [crush] Selecting the current rack

2015-11-24 Thread Gregory Farnum
On Tue, Nov 24, 2015 at 1:37 PM, Wido den Hollander wrote: > On 11/24/2015 07:00 PM, Emmanuel Lacour wrote: >> >> Dear ceph users, >> >> >> I try to write a crush ruleset that will, for a pool size of 3, put a >> copy in another host in the local rack and a copy in another rack. I now >> how to do

Re: [ceph-users] Storing Metadata

2015-11-24 Thread Gregory Farnum
On Tue, Nov 24, 2015 at 1:50 PM, James Gallagher wrote: > Hi there, > > I'm currently following the Ceph QSGs and have currently finished the > Storage Cluster Quick Start and have the current topology of > > admin-node - node1 (mon, mds) > - node2 (osd0) > - no

Re: [ceph-users] MDS memory usage

2015-11-24 Thread Gregory Farnum
On Tue, Nov 24, 2015 at 10:26 PM, Mike Miller wrote: > Hi, > > in my cluster with 16 OSD daemons and more than 20 million files on cephfs, > the memory usage on MDS is around 16 GB. It seems that 'mds cache size' has > no real influence on the memory usage of the MDS. > > Is there a formula that r

Re: [ceph-users] MDS memory usage

2015-11-25 Thread Gregory Farnum
> would probably help very much to balance the load. But I am afraid what > everybody says about stability issues. > > Is more than one MDS considered stable enough with hammer? You're correct, it's unfortunately not recommended yet. :( -Greg > > Thanks and regards

Re: [ceph-users] network failover with public/custer network - is that possible

2015-11-25 Thread Gregory Farnum
On Wed, Nov 25, 2015 at 8:37 AM, Götz Reinicke - IT Koordinator wrote: > Hi, > > discussing some design questions we came across the failover possibility > of cephs network configuration. > > If I just have a public network, all traffic is crossing that lan. > > With public and cluster network I c

Re: [ceph-users] Would HEALTH_DISASTER be a good addition?

2015-11-25 Thread Gregory Farnum
On Wed, Nov 25, 2015 at 11:09 AM, Wido den Hollander wrote: > Hi, > > Currently we have OK, WARN and ERR as states for a Ceph cluster. > > Now, it could happen that while a Ceph cluster is in WARN state certain > PGs are not available due to being in peering or any non-active+? state. > > When mon

Re: [ceph-users] Modification Time of RBD Images

2015-11-26 Thread Gregory Farnum
I don't think anything tracks this explicitly for RBD, but each RADOS object does maintain an mtime you can check via the rados tool. You could write a script to iterate through all the objects in the image and find the most recent mtime (although a custom librados binary will be faster if you want

Re: [ceph-users] Global, Synchronous Blocked Requests

2015-11-27 Thread Gregory Farnum
On Fri, Nov 27, 2015 at 10:24 PM, Brian Felton wrote: > Greetings Ceph Community, > > We are running a Hammer cluster (0.94.3-1) in production that recently > experienced asymptotic performance degradation. We've been migrating data > from an older non-Ceph cluster at a fairly steady pace for the

Re: [ceph-users] rbd_inst.create

2015-11-30 Thread Gregory Farnum
On Nov 27, 2015 3:34 AM, "NEVEU Stephane" wrote: > > Ok, I think I got it. It seems to come from here : > > tracker.ceph.com/issues/6047 > > > > I’m trying to snapshot an image while I previously made a snapshot of my pool… whereas it just works fine when using a brand new pool. I’m using ceph v0.

Re: [ceph-users] CRUSH Algorithm

2015-11-30 Thread Gregory Farnum
The code is in ceph/src/crush of the gut repo, but it's pretty opaque. If you go to the Ceph site and look through the pages there's one about "publications" (or maybe just documentation? I think publications) that hosts a paper on how CRUSH works. IIRC it's using the jenkins hash on the object na

Re: [ceph-users] how to mount a bootable VM image file?

2015-12-02 Thread Gregory Farnum
On Wednesday, December 2, 2015, Judd Maltin wrote: > I'm using OpenStack to create VMs. They're KVM VMs, and I can see all the > authentication information I need on the process tree. I want to mount > this bootable image on the hypervizor node to access its filesystem and fix > a file I messed

Re: [ceph-users] Flapping OSDs, Large meta directories in OSDs

2015-12-02 Thread Gregory Farnum
On Tue, Dec 1, 2015 at 10:02 AM, Tom Christensen wrote: > Another thing that we don't quite grasp is that when we see slow requests > now they almost always, probably 95% have the "known_if_redirected" state > set. What does this state mean? Does it indicate we have OSD maps that are > lagging a

Re: [ceph-users] OSD crash, unable to restart

2015-12-02 Thread Gregory Farnum
On Wed, Dec 2, 2015 at 10:54 AM, Major Csaba wrote: > Hi, > > I have a small cluster(5 nodes, 20OSDs), where an OSD crashed. There is no > any other signal of problems. No kernel message, so the disks seem to be OK. > > I tried to restart the OSD but the process stops almost immediately with the >

Re: [ceph-users] OSD crash, unable to restart

2015-12-02 Thread Gregory Farnum
On Wed, Dec 2, 2015 at 11:11 AM, Major Csaba wrote: > Hi, > [ sorry, I accidentaly left out the list address ] > > This is the content of the LOG file in the directory > /var/lib/ceph/osd/ceph-7/current/omap: > 2015/12/02-18:48:12.241386 7f805fc27900 Recovering log #26281 > 2015/12/02-18:48:12.242

Re: [ceph-users] osd process threads stack up on osds failure

2015-12-07 Thread Gregory Farnum
On Mon, Dec 7, 2015 at 6:59 AM, Kostis Fardelas wrote: > Hi cephers, > after one OSD node crash (6 OSDs in total), we experienced an increase > of approximately 230-260 threads for every other OSD node. We have 26 > OSD nodes with 6 OSDs per node, so this is approximately 40 threads > per osd. The

Re: [ceph-users] [Ceph-Users] Upgrade Path to Hammer

2015-12-07 Thread Gregory Farnum
As that ticket indicates, older versions of the code didn't create the backtraces, so obviously they aren't present. That certainly includes Dumpling! -Greg On Monday, December 7, 2015, Shinobu Kinjo wrote: > Hello, > > Have any of you tried to upgrade the Ceph cluster through the following > up

Re: [ceph-users] [Ceph-Users] Upgrade Path to Hammer

2015-12-07 Thread Gregory Farnum
... > > Shinobu > > - Original Message - > From: "Gregory Farnum" > > To: "Shinobu Kinjo" > > Cc: "ceph-users" > > Sent: Tuesday, December 8, 2015 10:36:34 AM > Subject: Re: [ceph-users] [Ceph-Users] Upgrade Path to Hammer > >

Re: [ceph-users] Ceph cache tier, flushed objects does not appear to be written on disk

2016-04-26 Thread Gregory Farnum
gt; 10c1ccc.__head_D073__1 > > -rw-r--r--. 1 root root 0 Mar 17 17:23 > 10fc00c.__head_8302__1 > > -rw-r--r--. 1 root root 0 Mar 17 17:23 > 10fcae7.__head_C3A7__1 > > [...] > > > > 2016-04-25 17:27 GMT+02:00 Grego

Re: [ceph-users] mount -t ceph

2016-04-27 Thread Gregory Farnum
On Wed, Apr 27, 2016 at 2:55 PM, Deneau, Tom wrote: > What kernel versions are required to be able to use CephFS thru mount -t ceph? The CephFS kernel client has been in for ages (2.6.34, I think?), but you want the absolute latest you can make happen if you're going to try it out. The actual mou

Re: [ceph-users] mount -t ceph

2016-04-27 Thread Gregory Farnum
n your distro kernel, not upstream. Afraid I don't know what goes on in...well, any of them, but certainly not SLES. ;) -Greg > > -- Tom > > >> -Original Message- >> From: Gregory Farnum [mailto:gfar...@redhat.com] >> Sent: Wednesday, April 27, 2016 2:59 P

Re: [ceph-users] Mapping RBD On Ceph Cluster Node

2016-04-29 Thread Gregory Farnum
On Friday, April 29, 2016, Edward Huyer wrote: > This is more of a "why" than a "can I/should I" question. > > The Ceph block device quickstart says (if I interpret it correctly) not to > use a physical machine as both a Ceph RBD client and a node for hosting > OSDs or other Ceph services. > > Is

Re: [ceph-users] Disabling POSIX locking semantics for CephFS

2016-05-03 Thread Gregory Farnum
On Tue, May 3, 2016 at 9:30 AM, Burkhard Linke wrote: > Hi, > > we have a number of legacy applications that do not cope well with the POSIX > locking semantics in CephFS due to missing locking support (e.g. flock > syscalls). We are able to fix some of these applications, but others are > binary

Re: [ceph-users] jewel, cephfs and selinux

2016-05-03 Thread Gregory Farnum
On Sun, May 1, 2016 at 5:52 PM, Andrus, Brian Contractor wrote: > All, > > > > I thought there was a way to mount CephFS using the kernel driver and be > able to honor selinux labeling. > > Right now, if I do ‘ls -lZ' on a mounted cephfs, I get question marks > instead of any contexts. > > When I

Re: [ceph-users] ceph degraded writes

2016-05-03 Thread Gregory Farnum
On Tue, May 3, 2016 at 4:10 PM, Ben Hines wrote: > The Hammer .93 to .94 notes said: > If upgrading from v0.93, setosd enable degraded writes = false on all osds > prior to upgrading. The degraded writes feature has been reverted due to > 11155. > > Our cluster is now on Infernalis 9.2.1 and we

Re: [ceph-users] hammer - lost object after just one OSD failure?

2016-05-04 Thread Gregory Farnum
On Wed, May 4, 2016 at 12:00 AM, Nikola Ciprich wrote: > Hi, > > I was doing some performance tuning on test cluster of just 2 > nodes (each 10 OSDs). I have test pool of 2 replicas (size=2, min_size=2) > > then one of OSD crashed due to failing harddrive. All remaining OSDs were > fine, but healt

Re: [ceph-users] Disabling POSIX locking semantics for CephFS

2016-05-04 Thread Gregory Farnum
On Wed, May 4, 2016 at 2:16 AM, Yan, Zheng wrote: > On Wed, May 4, 2016 at 4:51 PM, Burkhard Linke > wrote: >> Hi, >> >> >> How does CephFS handle locking in case of missing explicit locking control >> (e.g. flock / fcntl)? And what's the default of mmap'ed memory access in >> that case? >> > > N

Re: [ceph-users] ACL support in Jewel using fuse and SAMBA

2016-05-06 Thread Gregory Farnum
On Fri, May 6, 2016 at 9:53 AM, Eric Eastman wrote: > I was doing some SAMBA testing and noticed that a kernel mounted share > acted differently then a fuse mounted share with Windows security on > my windows client. I cut my test down to as simple as possible, and I > am seeing the kernel mounted

Re: [ceph-users] CephFS + CTDB/Samba - MDS session timeout on lockfile

2016-05-09 Thread Gregory Farnum
On Mon, May 9, 2016 at 8:48 AM, Sage Weil wrote: > On Mon, 9 May 2016, Nick Fisk wrote: >> Hi All, >> >> I've been testing an active/active Samba cluster over CephFS, performance >> seems really good with small files compared to Gluster. Soft reboots work >> beautifully with little to no interrupt

Re: [ceph-users] ACL support in Jewel using fuse and SAMBA

2016-05-09 Thread Gregory Farnum
On Sat, May 7, 2016 at 9:53 PM, Eric Eastman wrote: > On Fri, May 6, 2016 at 2:14 PM, Eric Eastman > wrote: > >> As it should be working, I will increase the logging level in my >> smb.conf file and see what info I can get out of the logs, and report back. > > Setting the log level = 20 in my smb

Re: [ceph-users] Mixed versions of Ceph Cluster and RadosGW

2016-05-11 Thread Gregory Farnum
Sadly not. RGW generally requires updates to the OSD-side object class code for a lot of its functionality andisnt expected to work against older clusters. :( On Wednesday, May 11, 2016, Saverio Proto wrote: > Hello, > > I have a production Ceph cluster running the latest Hammer Release. > > We

Re: [ceph-users] wrong exit status if bucket already exists

2016-05-11 Thread Gregory Farnum
Yes, it's intentional. All ceph CLI operations are idempotent. On Tuesday, May 10, 2016, Swapnil Jain wrote: > Hi > > I am using infernalis 9.2.1. While creating bucket, if the bucket already > exists, its still returns 0 as exit status. Is it intentional out of some > reason or a bug? > > > > r

Re: [ceph-users] How do ceph clients determine a monitor's address (and esp. port) for initial connection?

2016-05-12 Thread Gregory Farnum
On Thu, May 12, 2016 at 6:45 AM, Christian Sarrasin wrote: > I'm trying to run monitors on a non-standard port and having trouble > connecting to them. The below shows the ceph client attempting to connect > to default port 6789 rather than 6788: > > ceph --cluster testcluster status > 2016-05-12

Re: [ceph-users] How do ceph clients determine a monitor's address (and esp. port) for initial connection?

2016-05-12 Thread Gregory Farnum
ot remembering how the parsing works for these, and it's a bit messy. You may be stuck using the full IP:port instead of host names for the "mon host" config, if it's not working without that. :/ -Greg > > Cheers, > Christian > > > On 12/05/16 21:06, Gregory Fa

Re: [ceph-users] Ceph ANT task and file is empty

2016-05-12 Thread Gregory Farnum
Can you provide more details about exactly what you're doing, and exactly how it fails? -Greg On Thu, May 12, 2016 at 12:49 AM, gjprabu wrote: > > Hi > > Anybody facing similar issue. Please share the solution. > > Regards > Prabu GJ > > On Wed, 11 May 2016 17:38:15 +0530 gjprabu wrote

Re: [ceph-users] CRUSH map help

2016-05-12 Thread Gregory Farnum
On Thu, May 12, 2016 at 2:36 PM, Stephen Mercier < stephen.merc...@attainia.com> wrote: > I'm trying to setup a crush rule, and I was hoping you guys could clarify > something for me. > > I have 4 storage nodes across 2 cabinets. (2x2) > > I have the crush hierarchy setup to reflect this layout (a

Re: [ceph-users] CRUSH map help

2016-05-12 Thread Gregory Farnum
On Thu, May 12, 2016 at 2:54 PM, Stephen Mercier < stephen.merc...@attainia.com> wrote: > Thank you very much for the thorough explanation. What you described was > one of the ways I was interpreting this. > > Now, out of curiosity, if I did: > > rule replicated_rack { > ruleset 0 > type replicate

Re: [ceph-users] Mount Jewel CephFS to CentOS6

2016-05-13 Thread Gregory Farnum
On Friday, May 13, 2016, Andrus, Brian Contractor wrote: > So I see that support for RHEL6 and derivatives was dropped in Jewel ( > http://ceph.com/releases/v10-2-0-jewel-released/) > > > > But is there backward compatibility to mount it using hammer on a node? > Doesn’t seem to be and that makes

Re: [ceph-users] How do ceph clients determine a monitor's address (and esp. port) for initial connection?

2016-05-13 Thread Gregory Farnum
On Fri, May 13, 2016 at 12:51 AM, Christian Sarrasin wrote: > Hi Greg, > > Thanks again and good guess! Amending testcluster.conf as follows: > > mon host = 192.168.10.201:6788 > mon addr = 192.168.10.201:6788 > > ... gets around the problem. > > having "mon host = mona:6788" also works. > > Shou

Re: [ceph-users] straw_calc_version

2016-05-13 Thread Gregory Farnum
On Fri, May 13, 2016 at 5:02 PM, Tu Holmes wrote: > Hello again Cephers... As I'm learning more and breaking more things, I'm > finding more things I don't know. > > So currently, with all of the other things since I started upgrading to > Jewel, I'm seeing this in my logs. > > crush map has straw

Re: [ceph-users] CEPH/CEPHFS upgrade questions (9.2.0 ---> 10.2.1)

2016-05-24 Thread Gregory Farnum
On Wed, May 18, 2016 at 6:04 PM, Goncalo Borges wrote: > Dear All... > > Our infrastructure is the following: > > - We use CEPH/CEPHFS (9.2.0) > - We have 3 mons and 8 storage servers supporting 8 OSDs each. > - We use SSDs for journals (2 SSDs per storage server, each serving 4 OSDs). > - We have

Re: [ceph-users] Jewel CephFS quota (setfattr, getfattr)

2016-05-24 Thread Gregory Farnum
On Mon, May 23, 2016 at 12:52 AM, Yan, Zheng wrote: > To enable quota, you need to pass "--client-quota" option to ceph-fuse Yeah, this is a bit tricky since the kernel just doesn't respect quota at all. Perhaps once the kernel does support them we should make this the default. Or do something li

Re: [ceph-users] Blocked ops, OSD consuming memory, hammer

2016-05-24 Thread Gregory Farnum
On Tue, May 24, 2016 at 2:16 PM, Heath Albritton wrote: > Having some problems with my cluster. Wondering if I could get some > troubleshooting tips: > > Running hammer 0.94.5. Small cluster with cache tiering. 3 spinning > nodes and 3 SSD nodes. > > Lots of blocked ops. OSDs are consuming the

Re: [ceph-users] CEPH/CEPHFS upgrade questions (9.2.0 ---> 10.2.1)

2016-05-25 Thread Gregory Farnum
x86_64.rpm > librbd1-10.2.1-0.el7.x86_64.rpm > librbd1-devel-10.2.1-0.el7.x86_64.rpm > librgw2-10.2.1-0.el7.x86_64.rpm > librgw2-devel-10.2.1-0.el7.x86_64.rpm > python-ceph-compat-10.2.1-0.el7.x86_64.rpm > python-cephfs-10.2.1-0.el7.x86_64.rpm > pyt

Re: [ceph-users] Blocked ops, OSD consuming memory, hammer

2016-05-25 Thread Gregory Farnum
On Tue, May 24, 2016 at 11:19 PM, Heath Albritton wrote: > Not going to attempt threading and apologies for the two messages on > the same topic. Christian is right, though. 3 nodes per tier, 8 SSDs > per node in the cache tier, 12 spinning disks in the cold tier. 10GE > client network with a s

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
On Fri, May 27, 2016 at 9:44 AM, Stillwell, Bryan J wrote: > I have a Ceph cluster at home that I¹ve been running CephFS on for the > last few years. Recently my MDS server became damaged and while > attempting to fix it I believe I¹ve destroyed by CephFS journal based off > this: > > 2016-05-25

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
On Fri, May 27, 2016 at 1:54 PM, Stillwell, Bryan J wrote: > On 5/27/16, 11:27 AM, "Gregory Farnum" wrote: > >>On Fri, May 27, 2016 at 9:44 AM, Stillwell, Bryan J >> wrote: >>> I have a Ceph cluster at home that I¹ve been running CephFS on for the >&g

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
What's the current full output of "ceph -s"? If you already had your MDS in damaged state, you might just need to mark it as repaired. That's a monitor command. On Fri, May 27, 2016 at 2:09 PM, Stillwell, Bryan J wrote: > On 5/27/16, 3:01 PM, "Gregory Farnum" w

Re: [ceph-users] Rebuilding/recreating CephFS journal?

2016-05-27 Thread Gregory Farnum
sortbitwise > pgmap v10096597: 480 pgs, 4 pools, 23718 GB data, 5951 kobjects > 35758 GB used, 11358 GB / 47116 GB avail > 479 active+clean >1 active+clean+scrubbing+deep Yeah, you should just need to mark mds 0 as repaired at this

Re: [ceph-users] Infernalis => Jewel: ceph-fuse regression concerning the automatic mount at boot?

2016-06-01 Thread Gregory Farnum
On Wed, Jun 1, 2016 at 10:23 AM, Francois Lafont wrote: > Hi, > > I have a Jewel Ceph cluster in OK state and I have a "ceph-fuse" Ubuntu > Trusty client with ceph Infernalis. The cephfs is mounted automatically > and perfectly during the boot via ceph-fuse and this line in /etc/fstab : > > ~# gre

Re: [ceph-users] Crashing OSDs (suicide timeout, following a single pool)

2016-06-01 Thread Gregory Farnum
If that pool is your metadata pool, it looks at a quick glance like it's timing out somewhere while reading and building up the omap contents (ie, the contents of a directory). Which might make sense if, say, you have very fragmented leveldb stores combined with very large CephFS directories. Tryin

Re: [ceph-users] Crashing OSDs (suicide timeout, following a single pool)

2016-06-01 Thread Gregory Farnum
/import thing to get the PG to its new locations, but just deleting it certainly won't help! -Greg > > Thanks, > Adam > > On Wed, Jun 1, 2016 at 4:33 PM, Gregory Farnum wrote: >> If that pool is your metadata pool, it looks at a quick glance like >> it's timing

Re: [ceph-users] Retrieve mds sessions

2016-06-02 Thread Gregory Farnum
On Thu, Jun 2, 2016 at 6:27 AM, Antonios Matsoukas wrote: > I am trying to get a list of the sessions from the MDS but i receive the > following error: > > root@monitor1:/var/log/ceph# ceph daemon mds.storage1 session ls > admin_socket: exception getting command descriptions: [Errno 2] No such fi

Re: [ceph-users] Crashing OSDs (suicide timeout, following a single pool)

2016-06-02 Thread Gregory Farnum
similar to the recovery process described in this post from >> 04/09/2015: >> http://ceph-users.ceph.narkive.com/lwDkR2fZ/recovering-incomplete-pgs-with-ceph-objectstore-tool >> Hopefully it works in your case too and you can the cluster back to a state >> that you can

Re: [ceph-users] CephFS in the wild

2016-06-05 Thread Gregory Farnum
On Wed, Jun 1, 2016 at 1:50 PM, Brady Deetz wrote: > Question: > I'm curious if there is anybody else out there running CephFS at the scale > I'm planning for. I'd like to know some of the issues you didn't expect that > I should be looking out for. I'd also like to simply see when CephFS hasn't >

Re: [ceph-users] ceph-fuse, fio largely better after migration Infernalis to Jewel, is my bench relevant?

2016-06-06 Thread Gregory Farnum
We had several metadata caching improvements in ceph-fuse recently which I think went in after Infernalis. That could explain it. -Greg On Monday, June 6, 2016, Francois Lafont wrote: > Hi, > > I have a little Ceph cluster in production with 5 cluster nodes and 2 > client nodes. The clients are

Re: [ceph-users] Difference between step choose and step chooseleaf

2016-06-08 Thread Gregory Farnum
On Wed, Jun 8, 2016 at 8:22 AM, George Shuklin wrote: > Hello. > > Can someone help me to see difference between step choose and step > chooseleaf in CRUSH map? When you run "choose" on a CRUSH bucket type, it selects CRUSH bucket nodes of that type. If you run chooseleaf, it selects leaf nodes u

Re: [ceph-users] "mount error 5 = Input/output error" with the CephFS file system from client node

2016-06-14 Thread Gregory Farnum
On Tue, Jun 14, 2016 at 4:29 AM, Rakesh Parkiti wrote: > Hello, > > Unable to mount the CephFS file system from client node with "mount error 5 > = Input/output error" > MDS was installed on a separate node. Ceph Cluster health is OK and mds > services are running. firewall was disabled across all

Re: [ceph-users] Ceph file change monitor

2016-06-16 Thread Gregory Farnum
On Wed, Jun 15, 2016 at 5:19 AM, siva kumar <85s...@gmail.com> wrote: > Yes , We need to similar to inotify/fanotity . > > came through link > http://docs.ceph.com/docs/master/dev/osd_internals/watch_notify/?highlight=notify#watch-notify > > Just want to know if i can use this ? > > If yes means ho

Re: [ceph-users] OSDs not coming up on one host

2016-06-16 Thread Gregory Farnum
e monitor logs before anybody can help. If they aren't, you are skipping steps and need to include OSD logs and things. ;) -Greg > > Best regards, > Kostis > > On 17 April 2015 at 02:04, Gregory Farnum wrote: >> The monitor looks like it's not generating a new

Re: [ceph-users] Ceph Day Switzerland slides and video

2016-06-16 Thread Gregory Farnum
On Wed, Jun 15, 2016 at 11:30 AM, Dan van der Ster wrote: > Dear Ceph Community, > > Yesterday we had the pleasure of hosting Ceph Day Switzerland, and we > wanted to let you know that the slides and videos of most talks have > been posted online: > > https://indico.cern.ch/event/542464/timetabl

Re: [ceph-users] Mysterious cache-tier flushing behavior

2016-06-17 Thread Gregory Farnum
Sounds like you've got deleted objects in the cache tier getting flushed (i.e., deleted) in the base tier. -Greg On Thursday, June 16, 2016, Christian Balzer wrote: > > Hello devs and other sage(sic) people, > > Ceph 0.94.5, cache tier in writeback mode. > > As mentioned before, I'm running a cr

Re: [ceph-users] Mysterious cache-tier flushing behavior

2016-06-17 Thread Gregory Farnum
; I can assure you, in that cluster objects don't get deleted. > > Christian > > On Fri, 17 Jun 2016 08:57:31 -0700 Gregory Farnum wrote: > >> Sounds like you've got deleted objects in the cache tier getting flushed >> (i.e., deleted) in the base tier. &g

<    8   9   10   11   12   13   14   15   16   17   >