On Fri, Oct 23, 2015 at 10:14 AM, Edward Ned Harvey (ceph)
wrote:
> Trying to figure out if ceph supports inotify, or some form of notification,
> I see this issue from 4 years ago:
>
> http://tracker.ceph.com/issues/1296
>
>
>
> And the corresponding discussion thread
>
> http://comments.gmane.or
You might see if http://tracker.ceph.com/issues/13060 could apply to
your cluster. If so upgrading to .94.4 should fix it.
*Don't* reset your OSD journal. That is never the answer and is
basically the same as trashing the OSD in question.
-Greg
On Tue, Oct 27, 2015 at 9:59 AM, Laurent GUERBY wro
On Thu, Oct 22, 2015 at 3:58 PM, Andras Pataki
wrote:
> Hi ceph users,
>
> We’ve upgraded to 0.94.4 (all ceph daemons got restarted) – and are in the
> middle of doing some rebalancing due to crush changes (removing some disks).
> During the rebalance, I see that some placement groups get stuck in
On Tue, Oct 27, 2015 at 11:03 AM, Gregory Farnum wrote:
> On Thu, Oct 22, 2015 at 3:58 PM, Andras Pataki
> wrote:
>> Hi ceph users,
>>
>> We’ve upgraded to 0.94.4 (all ceph daemons got restarted) – and are in the
>> middle of doing some rebalancing due to crus
On Tue, Oct 27, 2015 at 11:22 AM, Andras Pataki
wrote:
> Hi Greg,
>
> No, unfortunately I haven¹t found any resolution to it. We are using
> cephfs, the whole installation is on 0.94.4. What I did notice is that
> performance is extremely poor when backfilling is happening. I wonder if
> timeou
[ Removed ceph-devel ]
On Wednesday, October 28, 2015, Libin Wu wrote:
> Hi, all
>
> As my understand, command "ceph daemon osd.x perf dump objecters" should
> output the perf data of osdc(librados). But when i use this command,
> why all those values are zero expcept map_epoch and map_inc. Foll
On Wed, Oct 28, 2015 at 8:38 PM, Yan, Zheng wrote:
> On Thu, Oct 29, 2015 at 1:10 AM, Burkhard Linke
>> I tried to dig into the ceph-fuse code, but I was unable to find the
>> fragment that is responsible for flushing the data from the page cache.
>>
>
> fuse kernel code invalidates page cache on
You can also extend that command line to specify specific block and
total sizes. Check the help text. :)
-Greg
On Thursday, October 29, 2015, Lindsay Mathieson <
lindsay.mathie...@gmail.com> wrote:
>
> On 29 October 2015 at 19:24, Burkhard Linke <
> burkhard.li...@computational.bio.uni-giessen.de
On Fri, Oct 30, 2015 at 6:20 PM, Artie Ziff wrote:
> Hello,
>
> In the RELEASE INFORMATION section of the hammer v0.94.3 issue tracker [1]
> the git commit SHA1 is: b2503b0e15c0b13f480f0835060479717b9cf935
>
> On the github page for Ceph Release v0.94.3 [2], when I click on the
> "95cefea" link [3
On Friday, October 30, 2015, mad Engineer wrote:
> i am learning ceph,block storage and read that each object size is 4 Mb.I
> am not clear about the concepts of object storage still what will happen if
> the actual size of data written to block is less than 4 Mb lets say 1
> Mb.Will it still cre
Regardless of what the crush tool does, I wouldn't muck around with the IDs
of the OSDs. The rest of Celh will probably not handle it well if the crush
IDs don't match the OSD numbers.
-Greg
On Monday, November 2, 2015, Loris Cuoghi wrote:
> Le 02/11/2015 12:47, Wido den Hollander a écrit :
>
>>
, and testing the two
> maps with :
>
> crushtool -i crush.map --test --show-statistics --rule 0 --num-rep 3 --min-x
> 1 --max-x $N --show-mappings
>
> (with $N varying from as little as 32 to "big numbers"TM) shows that nearly
> the 50% of the mappings changed, in a 10 host
When you have a caching pool in writeback mode, updates to objects
(including deletes) are handled by writeback rather than writethrough.
Since there's no other activity against these pools, there is nothing
prompting the cache pool to flush updates out to the backing pool, so
the backing pool hasn
11, Robert LeBlanc > написал(а):
> >
> > -----BEGIN PGP SIGNED MESSAGE-
> > Hash: SHA256
> >
> > Try:
> >
> > rados -p {cachepool} cache-flush-evict-all
> >
> > and see if the objects clean up.
> > -
>
It shouldn't be -- if you changed pg_num then a bunch of PGs will need to
move and will report in this state. We can check more thoroughly if you
provide the full "Ceph -s" output. (Stuff to check for: that all PGs are
active, none are degraded, etc)
-Greg
On Wednesday, November 4, 2015, Erming Pe
On Wed, Nov 4, 2015 at 12:27 PM, Rick Balsano wrote:
> Just following up since this thread went silent after a few comments showing
> similar concerns, but no explanation of the behavior. Can anyone point to
> some code or documentation which explains how to estimate the expected
> number of TCP c
On Fri, Nov 6, 2015 at 2:03 AM, Wido den Hollander wrote:
> Hi,
>
> Since Ceph Hammer we can protect pools from being removed from the
> cluster, but we can't protect against this:
>
> $ rbd ls|xargs -n 1 rbd rm
>
> That would remove all not opened RBD images from the cluster.
>
> This requires di
http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshooting-pg/
:)
On Friday, November 6, 2015, Philipp Schwaha wrote:
> Hi,
>
> I have an issue with my (small) ceph cluster after an osd failed.
> ceph -s reports the following:
> cluster 2752438a-a33e-4df4-b9ec-beae32d00aad
>
With that release it shouldn't be the EC pool causing trouble; it's the
CRUSH tunables also mentioned in that thread. Instructions should be
available in the docs for using older tunable that are compatible with
kernel 3.13.
-Greg
On Saturday, November 7, 2015, Bogdan SOLGA wrote:
> Hello, every
On Mon, Nov 9, 2015 at 6:57 AM, Burkhard Linke
wrote:
> Hi,
>
> On 11/09/2015 02:07 PM, Burkhard Linke wrote:
>>
>> Hi,
>
> *snipsnap*
>
>>
>>
>> Cluster is running Hammer 0.94.5 on top of Ubuntu 14.04. Clients use
>> ceph-fuse with patches for improved page cache handling, but the problem
>> also
The daemons print this in their debug logs on every boot. (There might
be a minimum debug level required, but I think it's at 0!)
-Greg
On Mon, Nov 9, 2015 at 7:23 AM, Wido den Hollander wrote:
> Hi,
>
> Recently I got my hands on a Ceph cluster which was pretty damaged due
> to a human error.
>
On Mon, Nov 9, 2015 at 9:42 AM, Deneau, Tom wrote:
> I don't have much experience with crush rules but wanted one that does the
> following:
>
> On a 3-node cluster, I wanted a rule where I could have an erasure-coded pool
> of k=3,m=2
> and where the first 3 chunks (the read chunks) are all on
Can you dump the metadata ops in flight on each ceph-fuse when it hangs?
ceph daemon mds_requests
-Greg
On Mon, Nov 9, 2015 at 8:06 AM, Burkhard Linke
wrote:
> Hi,
>
> On 11/09/2015 04:03 PM, Gregory Farnum wrote:
>>
>> On Mon, Nov 9, 2015 at 6:57 AM, Burkhard Linke
On Tue, Nov 10, 2015 at 6:32 AM, Oleksandr Natalenko
wrote:
> Hello.
>
> We have CephFS deployed over Ceph cluster (0.94.5).
>
> We experience constant MDS restarting under high IOPS workload (e.g.
> rsyncing lots of small mailboxes from another storage to CephFS using
> ceph-fuse client). First,
On Wednesday, November 11, 2015, Wido den Hollander wrote:
> On 11/10/2015 09:49 PM, Vickey Singh wrote:
> > On Mon, Nov 9, 2015 at 8:16 PM, Wido den Hollander > wrote:
> >
> >> On 11/09/2015 05:27 PM, Vickey Singh wrote:
> >>> Hello Ceph Geeks
> >>>
> >>> Need your comments with my understandin
On Wed, Nov 11, 2015 at 2:28 PM, Eric Eastman
wrote:
> On Wed, Nov 11, 2015 at 11:09 AM, John Spray wrote:
>> On Wed, Nov 11, 2015 at 5:39 PM, Eric Eastman
>> wrote:
>>> I am trying to figure out why my Ceph file system is not freeing
>>> space. Using Ceph 9.1.0 I created a file system with sna
I think I saw somebody working on a RADOS interface to Apache Hadoop once,
maybe search for that?
Your other option is to try and make use of object classes directly, but
that's a bit orimitive to build full map-reduce on top of without a lot of
effort.
-Greg
On Friday, November 13, 2015, min fang
Somebody else will need to do the diagnosis, but it'll help them if
you can get logs with "debug ms = 1", "debug osd = 20" in the log.
Based on the required features update in the crush map, it looks like
maybe you've upgraded some of your OSDs — is that a thing happening
right now? Perhaps you up
What's the full output of "Ceph -s"? Are your new crush rules actually
satisfiable? Is your cluster filling up?
-Greg
On Saturday, November 14, 2015, Peter Theobald wrote:
> Hi list,
>
> I have a 3 node ceph cluster with a total of 9 ods (2,3 and 4 with
> different size drives). I changed the la
On Monday, November 16, 2015, min fang wrote:
> Is this function used in detach rx buffer, and complete IO back to the
> caller? From the code, I think this function will not interact with OSD or
> MON side, which means, we just cancel IO from client side. Am I right?
>
> Thanks.
>
Right. If th
This usually means your crush mapping for the pool in question is
unsatisfiable. Check what the rule is doing.
-Greg
On Friday, November 20, 2015, German Anders wrote:
> Hi all, I've finished the install of a new ceph cluster with infernalis
> 9.2.0 release. But I'm getting the following error m
On Wed, Nov 18, 2015 at 4:34 AM, Межов Игорь Александрович
wrote:
> Hi!
>
> As for my previous message, digging mailing list gave me only one method to
> fix
> inconsistency - truncate object files in a filesystem to a size, that they
> have
> in ceph metadata:
>
> http://www.spinics.net/lists/c
On Thu, Nov 19, 2015 at 8:56 PM, Richard Gray wrote:
> Hi,
>
> Running 'health detail' on our Ceph cluster this morning, I notice a warning
> about one of the pools having significantly more objects per placement group
> than the cluster average.
>
> ceph> health detail
> HEALTH_WARN pool cas_back
On Mon, Nov 23, 2015 at 10:05 AM, German Anders wrote:
> Hi all,
>
> I want to know if there's any improvement or update regarding ceph 0.94.5
> with accelio, I've an already configured cluster (with no data on it) and I
> would like to know if there's a way to 'modify' the cluster in order to use
On Mon, Nov 23, 2015 at 8:44 AM, louis wrote:
> Hi, if I submit read or write io in a sequence from a ceph client, will
> these sequence will be kept in osds side? Thanks
Any writes from the same client, to the same object, will be ordered
with respect to one another. But there are no other guara
Yep. I think you can inject it into the running mons without
restarting as well (injectargs).
-Greg
On Mon, Nov 23, 2015 at 11:46 AM, Yujian Peng wrote:
> The mons in my production cluster have a very high cpu usage 100%.
> I think it may be caused by the leveldb compression.
> How yo disable lev
On Fri, Nov 20, 2015 at 11:33 AM, Simon Engelsman wrote:
> Hi,
>
> We've experienced a very weird problem last week with our Ceph
> cluster. We would like to ask your opinion(s) and advice
>
> Our dedicated Ceph OSD nodes run with:
>
> Total platform
> - IO Average: 2500 wrps, ~ 600 rps
> - Replic
Yeah, the write proxying is pretty new and the fact that it's missing from
an oddball like READFORWARD isn't surprising. (Not good, exactly, but not
surprising.) What are you doing with this caching mode?
On Thu, Nov 19, 2015 at 10:34 AM, Nick Fisk wrote:
> Don’t know why that URL got changed, i
On Tue, Nov 24, 2015 at 1:37 PM, Wido den Hollander wrote:
> On 11/24/2015 07:00 PM, Emmanuel Lacour wrote:
>>
>> Dear ceph users,
>>
>>
>> I try to write a crush ruleset that will, for a pool size of 3, put a
>> copy in another host in the local rack and a copy in another rack. I now
>> how to do
On Tue, Nov 24, 2015 at 1:50 PM, James Gallagher
wrote:
> Hi there,
>
> I'm currently following the Ceph QSGs and have currently finished the
> Storage Cluster Quick Start and have the current topology of
>
> admin-node - node1 (mon, mds)
> - node2 (osd0)
> - no
On Tue, Nov 24, 2015 at 10:26 PM, Mike Miller wrote:
> Hi,
>
> in my cluster with 16 OSD daemons and more than 20 million files on cephfs,
> the memory usage on MDS is around 16 GB. It seems that 'mds cache size' has
> no real influence on the memory usage of the MDS.
>
> Is there a formula that r
> would probably help very much to balance the load. But I am afraid what
> everybody says about stability issues.
>
> Is more than one MDS considered stable enough with hammer?
You're correct, it's unfortunately not recommended yet. :(
-Greg
>
> Thanks and regards
On Wed, Nov 25, 2015 at 8:37 AM, Götz Reinicke - IT Koordinator
wrote:
> Hi,
>
> discussing some design questions we came across the failover possibility
> of cephs network configuration.
>
> If I just have a public network, all traffic is crossing that lan.
>
> With public and cluster network I c
On Wed, Nov 25, 2015 at 11:09 AM, Wido den Hollander wrote:
> Hi,
>
> Currently we have OK, WARN and ERR as states for a Ceph cluster.
>
> Now, it could happen that while a Ceph cluster is in WARN state certain
> PGs are not available due to being in peering or any non-active+? state.
>
> When mon
I don't think anything tracks this explicitly for RBD, but each RADOS
object does maintain an mtime you can check via the rados tool. You could
write a script to iterate through all the objects in the image and find the
most recent mtime (although a custom librados binary will be faster if you
want
On Fri, Nov 27, 2015 at 10:24 PM, Brian Felton wrote:
> Greetings Ceph Community,
>
> We are running a Hammer cluster (0.94.3-1) in production that recently
> experienced asymptotic performance degradation. We've been migrating data
> from an older non-Ceph cluster at a fairly steady pace for the
On Nov 27, 2015 3:34 AM, "NEVEU Stephane"
wrote:
>
> Ok, I think I got it. It seems to come from here :
>
> tracker.ceph.com/issues/6047
>
>
>
> I’m trying to snapshot an image while I previously made a snapshot of my
pool… whereas it just works fine when using a brand new pool. I’m using
ceph v0.
The code is in ceph/src/crush of the gut repo, but it's pretty opaque. If
you go to the Ceph site and look through the pages there's one about
"publications" (or maybe just documentation? I think publications) that
hosts a paper on how CRUSH works.
IIRC it's using the jenkins hash on the object na
On Wednesday, December 2, 2015, Judd Maltin wrote:
> I'm using OpenStack to create VMs. They're KVM VMs, and I can see all the
> authentication information I need on the process tree. I want to mount
> this bootable image on the hypervizor node to access its filesystem and fix
> a file I messed
On Tue, Dec 1, 2015 at 10:02 AM, Tom Christensen wrote:
> Another thing that we don't quite grasp is that when we see slow requests
> now they almost always, probably 95% have the "known_if_redirected" state
> set. What does this state mean? Does it indicate we have OSD maps that are
> lagging a
On Wed, Dec 2, 2015 at 10:54 AM, Major Csaba wrote:
> Hi,
>
> I have a small cluster(5 nodes, 20OSDs), where an OSD crashed. There is no
> any other signal of problems. No kernel message, so the disks seem to be OK.
>
> I tried to restart the OSD but the process stops almost immediately with the
>
On Wed, Dec 2, 2015 at 11:11 AM, Major Csaba wrote:
> Hi,
> [ sorry, I accidentaly left out the list address ]
>
> This is the content of the LOG file in the directory
> /var/lib/ceph/osd/ceph-7/current/omap:
> 2015/12/02-18:48:12.241386 7f805fc27900 Recovering log #26281
> 2015/12/02-18:48:12.242
On Mon, Dec 7, 2015 at 6:59 AM, Kostis Fardelas wrote:
> Hi cephers,
> after one OSD node crash (6 OSDs in total), we experienced an increase
> of approximately 230-260 threads for every other OSD node. We have 26
> OSD nodes with 6 OSDs per node, so this is approximately 40 threads
> per osd. The
As that ticket indicates, older versions of the code didn't create the
backtraces, so obviously they aren't present. That certainly includes
Dumpling!
-Greg
On Monday, December 7, 2015, Shinobu Kinjo wrote:
> Hello,
>
> Have any of you tried to upgrade the Ceph cluster through the following
> up
...
>
> Shinobu
>
> - Original Message -
> From: "Gregory Farnum" >
> To: "Shinobu Kinjo" >
> Cc: "ceph-users" >
> Sent: Tuesday, December 8, 2015 10:36:34 AM
> Subject: Re: [ceph-users] [Ceph-Users] Upgrade Path to Hammer
>
>
gt; 10c1ccc.__head_D073__1
>
> -rw-r--r--. 1 root root 0 Mar 17 17:23
> 10fc00c.__head_8302__1
>
> -rw-r--r--. 1 root root 0 Mar 17 17:23
> 10fcae7.__head_C3A7__1
>
> [...]
>
>
>
> 2016-04-25 17:27 GMT+02:00 Grego
On Wed, Apr 27, 2016 at 2:55 PM, Deneau, Tom wrote:
> What kernel versions are required to be able to use CephFS thru mount -t ceph?
The CephFS kernel client has been in for ages (2.6.34, I think?), but
you want the absolute latest you can make happen if you're going to
try it out.
The actual mou
n your distro kernel, not upstream. Afraid I
don't know what goes on in...well, any of them, but certainly not
SLES. ;)
-Greg
>
> -- Tom
>
>
>> -Original Message-
>> From: Gregory Farnum [mailto:gfar...@redhat.com]
>> Sent: Wednesday, April 27, 2016 2:59 P
On Friday, April 29, 2016, Edward Huyer wrote:
> This is more of a "why" than a "can I/should I" question.
>
> The Ceph block device quickstart says (if I interpret it correctly) not to
> use a physical machine as both a Ceph RBD client and a node for hosting
> OSDs or other Ceph services.
>
> Is
On Tue, May 3, 2016 at 9:30 AM, Burkhard Linke
wrote:
> Hi,
>
> we have a number of legacy applications that do not cope well with the POSIX
> locking semantics in CephFS due to missing locking support (e.g. flock
> syscalls). We are able to fix some of these applications, but others are
> binary
On Sun, May 1, 2016 at 5:52 PM, Andrus, Brian Contractor
wrote:
> All,
>
>
>
> I thought there was a way to mount CephFS using the kernel driver and be
> able to honor selinux labeling.
>
> Right now, if I do ‘ls -lZ' on a mounted cephfs, I get question marks
> instead of any contexts.
>
> When I
On Tue, May 3, 2016 at 4:10 PM, Ben Hines wrote:
> The Hammer .93 to .94 notes said:
> If upgrading from v0.93, setosd enable degraded writes = false on all osds
> prior to upgrading. The degraded writes feature has been reverted due to
> 11155.
>
> Our cluster is now on Infernalis 9.2.1 and we
On Wed, May 4, 2016 at 12:00 AM, Nikola Ciprich
wrote:
> Hi,
>
> I was doing some performance tuning on test cluster of just 2
> nodes (each 10 OSDs). I have test pool of 2 replicas (size=2, min_size=2)
>
> then one of OSD crashed due to failing harddrive. All remaining OSDs were
> fine, but healt
On Wed, May 4, 2016 at 2:16 AM, Yan, Zheng wrote:
> On Wed, May 4, 2016 at 4:51 PM, Burkhard Linke
> wrote:
>> Hi,
>>
>>
>> How does CephFS handle locking in case of missing explicit locking control
>> (e.g. flock / fcntl)? And what's the default of mmap'ed memory access in
>> that case?
>>
>
> N
On Fri, May 6, 2016 at 9:53 AM, Eric Eastman
wrote:
> I was doing some SAMBA testing and noticed that a kernel mounted share
> acted differently then a fuse mounted share with Windows security on
> my windows client. I cut my test down to as simple as possible, and I
> am seeing the kernel mounted
On Mon, May 9, 2016 at 8:48 AM, Sage Weil wrote:
> On Mon, 9 May 2016, Nick Fisk wrote:
>> Hi All,
>>
>> I've been testing an active/active Samba cluster over CephFS, performance
>> seems really good with small files compared to Gluster. Soft reboots work
>> beautifully with little to no interrupt
On Sat, May 7, 2016 at 9:53 PM, Eric Eastman
wrote:
> On Fri, May 6, 2016 at 2:14 PM, Eric Eastman
> wrote:
>
>> As it should be working, I will increase the logging level in my
>> smb.conf file and see what info I can get out of the logs, and report back.
>
> Setting the log level = 20 in my smb
Sadly not. RGW generally requires updates to the OSD-side object class code
for a lot of its functionality andisnt expected to work against older
clusters. :(
On Wednesday, May 11, 2016, Saverio Proto wrote:
> Hello,
>
> I have a production Ceph cluster running the latest Hammer Release.
>
> We
Yes, it's intentional. All ceph CLI operations are idempotent.
On Tuesday, May 10, 2016, Swapnil Jain wrote:
> Hi
>
> I am using infernalis 9.2.1. While creating bucket, if the bucket already
> exists, its still returns 0 as exit status. Is it intentional out of some
> reason or a bug?
>
>
>
> r
On Thu, May 12, 2016 at 6:45 AM, Christian Sarrasin
wrote:
> I'm trying to run monitors on a non-standard port and having trouble
> connecting to them. The below shows the ceph client attempting to connect
> to default port 6789 rather than 6788:
>
> ceph --cluster testcluster status
> 2016-05-12
ot remembering how the parsing works for these, and
it's a bit messy. You may be stuck using the full IP:port instead of
host names for the "mon host" config, if it's not working without
that. :/
-Greg
>
> Cheers,
> Christian
>
>
> On 12/05/16 21:06, Gregory Fa
Can you provide more details about exactly what you're doing, and
exactly how it fails?
-Greg
On Thu, May 12, 2016 at 12:49 AM, gjprabu wrote:
>
> Hi
>
> Anybody facing similar issue. Please share the solution.
>
> Regards
> Prabu GJ
>
> On Wed, 11 May 2016 17:38:15 +0530 gjprabu wrote
On Thu, May 12, 2016 at 2:36 PM, Stephen Mercier <
stephen.merc...@attainia.com> wrote:
> I'm trying to setup a crush rule, and I was hoping you guys could clarify
> something for me.
>
> I have 4 storage nodes across 2 cabinets. (2x2)
>
> I have the crush hierarchy setup to reflect this layout (a
On Thu, May 12, 2016 at 2:54 PM, Stephen Mercier <
stephen.merc...@attainia.com> wrote:
> Thank you very much for the thorough explanation. What you described was
> one of the ways I was interpreting this.
>
> Now, out of curiosity, if I did:
>
> rule replicated_rack {
> ruleset 0
> type replicate
On Friday, May 13, 2016, Andrus, Brian Contractor wrote:
> So I see that support for RHEL6 and derivatives was dropped in Jewel (
> http://ceph.com/releases/v10-2-0-jewel-released/)
>
>
>
> But is there backward compatibility to mount it using hammer on a node?
> Doesn’t seem to be and that makes
On Fri, May 13, 2016 at 12:51 AM, Christian Sarrasin
wrote:
> Hi Greg,
>
> Thanks again and good guess! Amending testcluster.conf as follows:
>
> mon host = 192.168.10.201:6788
> mon addr = 192.168.10.201:6788
>
> ... gets around the problem.
>
> having "mon host = mona:6788" also works.
>
> Shou
On Fri, May 13, 2016 at 5:02 PM, Tu Holmes wrote:
> Hello again Cephers... As I'm learning more and breaking more things, I'm
> finding more things I don't know.
>
> So currently, with all of the other things since I started upgrading to
> Jewel, I'm seeing this in my logs.
>
> crush map has straw
On Wed, May 18, 2016 at 6:04 PM, Goncalo Borges
wrote:
> Dear All...
>
> Our infrastructure is the following:
>
> - We use CEPH/CEPHFS (9.2.0)
> - We have 3 mons and 8 storage servers supporting 8 OSDs each.
> - We use SSDs for journals (2 SSDs per storage server, each serving 4 OSDs).
> - We have
On Mon, May 23, 2016 at 12:52 AM, Yan, Zheng wrote:
> To enable quota, you need to pass "--client-quota" option to ceph-fuse
Yeah, this is a bit tricky since the kernel just doesn't respect quota
at all. Perhaps once the kernel does support them we should make this
the default. Or do something li
On Tue, May 24, 2016 at 2:16 PM, Heath Albritton wrote:
> Having some problems with my cluster. Wondering if I could get some
> troubleshooting tips:
>
> Running hammer 0.94.5. Small cluster with cache tiering. 3 spinning
> nodes and 3 SSD nodes.
>
> Lots of blocked ops. OSDs are consuming the
x86_64.rpm
> librbd1-10.2.1-0.el7.x86_64.rpm
> librbd1-devel-10.2.1-0.el7.x86_64.rpm
> librgw2-10.2.1-0.el7.x86_64.rpm
> librgw2-devel-10.2.1-0.el7.x86_64.rpm
> python-ceph-compat-10.2.1-0.el7.x86_64.rpm
> python-cephfs-10.2.1-0.el7.x86_64.rpm
> pyt
On Tue, May 24, 2016 at 11:19 PM, Heath Albritton wrote:
> Not going to attempt threading and apologies for the two messages on
> the same topic. Christian is right, though. 3 nodes per tier, 8 SSDs
> per node in the cache tier, 12 spinning disks in the cold tier. 10GE
> client network with a s
On Fri, May 27, 2016 at 9:44 AM, Stillwell, Bryan J
wrote:
> I have a Ceph cluster at home that I¹ve been running CephFS on for the
> last few years. Recently my MDS server became damaged and while
> attempting to fix it I believe I¹ve destroyed by CephFS journal based off
> this:
>
> 2016-05-25
On Fri, May 27, 2016 at 1:54 PM, Stillwell, Bryan J
wrote:
> On 5/27/16, 11:27 AM, "Gregory Farnum" wrote:
>
>>On Fri, May 27, 2016 at 9:44 AM, Stillwell, Bryan J
>> wrote:
>>> I have a Ceph cluster at home that I¹ve been running CephFS on for the
>&g
What's the current full output of "ceph -s"?
If you already had your MDS in damaged state, you might just need to
mark it as repaired. That's a monitor command.
On Fri, May 27, 2016 at 2:09 PM, Stillwell, Bryan J
wrote:
> On 5/27/16, 3:01 PM, "Gregory Farnum" w
sortbitwise
> pgmap v10096597: 480 pgs, 4 pools, 23718 GB data, 5951 kobjects
> 35758 GB used, 11358 GB / 47116 GB avail
> 479 active+clean
>1 active+clean+scrubbing+deep
Yeah, you should just need to mark mds 0 as repaired at this
On Wed, Jun 1, 2016 at 10:23 AM, Francois Lafont wrote:
> Hi,
>
> I have a Jewel Ceph cluster in OK state and I have a "ceph-fuse" Ubuntu
> Trusty client with ceph Infernalis. The cephfs is mounted automatically
> and perfectly during the boot via ceph-fuse and this line in /etc/fstab :
>
> ~# gre
If that pool is your metadata pool, it looks at a quick glance like
it's timing out somewhere while reading and building up the omap
contents (ie, the contents of a directory). Which might make sense if,
say, you have very fragmented leveldb stores combined with very large
CephFS directories. Tryin
/import thing to get the PG to its new
locations, but just deleting it certainly won't help!
-Greg
>
> Thanks,
> Adam
>
> On Wed, Jun 1, 2016 at 4:33 PM, Gregory Farnum wrote:
>> If that pool is your metadata pool, it looks at a quick glance like
>> it's timing
On Thu, Jun 2, 2016 at 6:27 AM, Antonios Matsoukas
wrote:
> I am trying to get a list of the sessions from the MDS but i receive the
> following error:
>
> root@monitor1:/var/log/ceph# ceph daemon mds.storage1 session ls
> admin_socket: exception getting command descriptions: [Errno 2] No such fi
similar to the recovery process described in this post from
>> 04/09/2015:
>> http://ceph-users.ceph.narkive.com/lwDkR2fZ/recovering-incomplete-pgs-with-ceph-objectstore-tool
>> Hopefully it works in your case too and you can the cluster back to a state
>> that you can
On Wed, Jun 1, 2016 at 1:50 PM, Brady Deetz wrote:
> Question:
> I'm curious if there is anybody else out there running CephFS at the scale
> I'm planning for. I'd like to know some of the issues you didn't expect that
> I should be looking out for. I'd also like to simply see when CephFS hasn't
>
We had several metadata caching improvements in ceph-fuse recently which I
think went in after Infernalis. That could explain it.
-Greg
On Monday, June 6, 2016, Francois Lafont wrote:
> Hi,
>
> I have a little Ceph cluster in production with 5 cluster nodes and 2
> client nodes. The clients are
On Wed, Jun 8, 2016 at 8:22 AM, George Shuklin wrote:
> Hello.
>
> Can someone help me to see difference between step choose and step
> chooseleaf in CRUSH map?
When you run "choose" on a CRUSH bucket type, it selects CRUSH bucket
nodes of that type. If you run chooseleaf, it selects leaf nodes
u
On Tue, Jun 14, 2016 at 4:29 AM, Rakesh Parkiti
wrote:
> Hello,
>
> Unable to mount the CephFS file system from client node with "mount error 5
> = Input/output error"
> MDS was installed on a separate node. Ceph Cluster health is OK and mds
> services are running. firewall was disabled across all
On Wed, Jun 15, 2016 at 5:19 AM, siva kumar <85s...@gmail.com> wrote:
> Yes , We need to similar to inotify/fanotity .
>
> came through link
> http://docs.ceph.com/docs/master/dev/osd_internals/watch_notify/?highlight=notify#watch-notify
>
> Just want to know if i can use this ?
>
> If yes means ho
e monitor logs before anybody can help. If they aren't,
you are skipping steps and need to include OSD logs and things. ;)
-Greg
>
> Best regards,
> Kostis
>
> On 17 April 2015 at 02:04, Gregory Farnum wrote:
>> The monitor looks like it's not generating a new
On Wed, Jun 15, 2016 at 11:30 AM, Dan van der Ster wrote:
> Dear Ceph Community,
>
> Yesterday we had the pleasure of hosting Ceph Day Switzerland, and we
> wanted to let you know that the slides and videos of most talks have
> been posted online:
>
> https://indico.cern.ch/event/542464/timetabl
Sounds like you've got deleted objects in the cache tier getting flushed
(i.e., deleted) in the base tier.
-Greg
On Thursday, June 16, 2016, Christian Balzer wrote:
>
> Hello devs and other sage(sic) people,
>
> Ceph 0.94.5, cache tier in writeback mode.
>
> As mentioned before, I'm running a cr
; I can assure you, in that cluster objects don't get deleted.
>
> Christian
>
> On Fri, 17 Jun 2016 08:57:31 -0700 Gregory Farnum wrote:
>
>> Sounds like you've got deleted objects in the cache tier getting flushed
>> (i.e., deleted) in the base tier.
&g
1201 - 1300 of 2358 matches
Mail list logo