Hi all,
We're reading from a Ceph Luminous pool using the librados asychronous I/O
API. We're seeing some concerning memory usage patterns when we read many
objects in sequence.
The expected behaviour is that our memory usage stabilises at a small
amount, since we're just fetching objects and ign
Quoting Yan, Zheng (uker...@gmail.com):
>
>
> please add '-f' option (trace child processes' syscall) to strace,
Good suggestion. We now see all apache child processes doing it's thing.
We have been, on and off, been stracing / debugging this issue. Nothing
obvious. We are still trying to get o
Hi,
Once in a while, today a bit more often, the MDS is logging the
following:
mds.mds1 [WRN] replayed op client.15327973:15585315,15585103 used ino
0x19918de but session next is 0x1873b8b
Nothing of importance is logged in the mds (debug_mds_log": "1/5").
What does this warning messag
On 09/12/2018 05:29 AM, Daniel Goldbach wrote:
Hi all,
We're reading from a Ceph Luminous pool using the librados asychronous
I/O API. We're seeing some concerning memory usage patterns when we
read many objects in sequence.
The expected behaviour is that our memory usage stabilises at a s
Yep, those completions are maintaining bufferlist references IIRC, so
they’re definitely holding the memory buffers in place!
On Wed, Sep 12, 2018 at 7:04 AM Casey Bodley wrote:
>
>
> On 09/12/2018 05:29 AM, Daniel Goldbach wrote:
> > Hi all,
> >
> > We're reading from a Ceph Luminous pool using
Is this osxfuse, the only and best performing way to mount a ceph
filesystem on an osx client?
http://docs.ceph.com/docs/mimic/dev/macos/
I am now testing cephfs performance on a client with the fio libaio
engine. This engine does not exist on osx, but there is a posixaio. Does
anyone have ex
We are benchmarking a test machine which has:
8 cores, 64GB RAM
12 * 12 TB HDD (SATA)
2 * 480 GB SSD (SATA)
1 * 240 GB SSD (NVME)
Ceph Mimic
Baseline benchmark for HDD only (Erasure Code 4+2)
Write 420 MB/s, 100 IOPS, 150ms latency
Read 1040 MB/s, 260 IOPS, 60ms latency
Now we moved WAL to the SS
Hi Jan,
how did you move the WAL and DB to the SSD/NVMe? By recreating the
OSDs or a different approach? Did you check afterwards that the
devices were really used for that purpose? We had to deal with that a
couple of months ago [1] and it's not really obvious if the new
devices are real
If you're writes are small enough (64k or smaller) they're being placed on
the WAL device regardless of where your DB is. If you change your testing
to use larger writes you should see a difference by adding the DB.
Please note that the community has never recommended using less than 120GB
DB for
Hi all,
I'm having trouble turning off the warning "1 pools have many more
objects per pg than average".
I've tried a lot of variations on the below, my current ceph.conf:
#...
[mon]
#...
mon_pg_warn_max_object_skew = 0
All of my monitors have been restarted.
Seems like I'm missing someth
The issue continues even when I do rados_aio_release(completion) at the end
of the readobj(..) definition in the example. Also, in our production code
we call rados_aio_release for each completion and we still see the issue
there. The release command doesn't guarantee instant release, so could it
b
[Moving this to ceph-users where it will get more eyeballs.]
On Wed, 12 Sep 2018, Andrew Cassera wrote:
> Hello,
>
> Any help would be appreciated. I just created two clusters in the lab. One
> cluster is running jewel 10.2.10 and the other cluster is running luminous
> 12.2.8. After creating
When having a hdd bluestore osd with collocated wal and db.
- What performance increase can be expected if one would move the wal to
an ssd?
- What performance increase can be expected if one would move the db to
an ssd?
- Would the performance be a lot if you have a very slow hdd (and thu
You already have a thread talking about benchmarking the addition of WAL
and DB partitions to an OSD. Why are you creating a new one about the
exact same thing? As with everything, the performance increase isn't even
solely answerable by which drives you have, there are a lot of factors that
coul
Sorry, I was wrong that it was you. I just double checked. But there is a
new thread as of this morning about this topic where someone is running
benchmark tests with numbers titled "Benchmark does not show gains with DB
on SSD".
On Wed, Sep 12, 2018 at 12:20 PM David Turner wrote:
> You alread
What thread? I have put this, with this specific subject so it is easier
to find in the future and this is not a 'sub question' of someone's
problem. Hoping for others to post their experience/results. I thought
if cern can give estimates, people here can to.
-Original Message-
From: D
That code bit is just "we have an incoming message with data", which is
what we'd expect, but means it's not very helpful for tracking down the
source of any leaks.
My guess is still very much that somehow there are deallocations missing
here. Internally, the synchronous API is wrapping the async
Eugene:
Between tests we destroyed the OSDs and created them from scratch. We used
Docker image to deploy Ceph on one machine.
I've seen that there are WAL/DB partitions created on the disks.
Should I also check somewhere in ceph config that it actually uses those?
David:
We used 4MB writes.
I kn
I couldn't find any sign of a networking issue at the OS or switches. No
changes have been made in those to get the cluster stable again. I
looked through a couple OSD logs and here is a selection of some of most
frequent errors they were getting. Maybe something below is more obvious
to you.
On 12/09/18 17:06, Ján Senko wrote:
We are benchmarking a test machine which has:
8 cores, 64GB RAM
12 * 12 TB HDD (SATA)
2 * 480 GB SSD (SATA)
1 * 240 GB SSD (NVME)
Ceph Mimic
Baseline benchmark for HDD only (Erasure Code 4+2)
Write 420 MB/s, 100 IOPS, 150ms latency
Read 1040 MB/s, 260 IOPS,
Did you restart the mons or inject the option?
Paul
2018-09-12 17:40 GMT+02:00 Chad William Seys :
> Hi all,
> I'm having trouble turning off the warning "1 pools have many more objects
> per pg than average".
>
> I've tried a lot of variations on the below, my current ceph.conf:
>
> #...
> [mo
On Tue, Sep 11, 2018 at 5:32 PM Benjamin Cherian
wrote:
> Ok, that’s good to know. I was planning on using an EC pool. Maybe I'll
> store some of the larger kv pairs in their own objects or move the metadata
> into it's own replicated pool entirely. If the storage mechanism is the
> same, is ther
Hi Paul,
Yes, all monitors have been restarted.
Chad.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Hi
replying to list - instead of directly by accident
---
Sorry I was camping for a week and was disconnected without data for the most
part
Yes it is over iSCSI - 2 iscsi nodes
We've set both iscsi and vmware hosts to SUSE recommended settings
in addition we've set round robi
Any chance you know the LBA or byte offset of the corruption so I can
compare it against the log?
On Wed, Sep 12, 2018 at 8:32 PM wrote:
>
> Hi Jason,
>
> On 2018-09-10 11:15:45-07:00 ceph-users wrote:
>
> On 2018-09-10 11:04:20-07:00 Jason Dillaman wrote:
>
>
> > In addition to this, we are se
On 2018-09-12 17:35:16-07:00 Jason Dillaman wrote:
Any chance you know the LBA or byte offset of the corruption so I can
compare it against the log?
The LBAs of the corruption are 0xA74F000 through 175435776
On Wed, Sep 12, 2018 at 8:32 PM wrote:
>
> Hi Jason,
On Wed, Sep 12, 2018 at 10:15 PM wrote:
>
> On 2018-09-12 17:35:16-07:00 Jason Dillaman wrote:
>
>
> Any chance you know the LBA or byte offset of the corruption so I can
> compare it against the log?
>
> The LBAs of the corruption are 0xA74F000 through 175435776
Are you saying the corruption sta
Greg, Paul,
Thank you for the feedback. This has been very enlightening. One last
question (for now at least). Are there any expected performance impacts
from having I/O to multiple pools from the same client? (Given how RGW and
CephFS store metadata, I would hope not, but I thought I'd ask.) Base
Nope there shouldn’t be any impact apart from the potential issues that
arise from breaking up the I/O stream. Which in the case of either a
saturated or mostly-idle RADOS cluster should not be an issue.
-Greg
On Wed, Sep 12, 2018 at 9:24 PM Benjamin Cherian
wrote:
> Greg, Paul,
>
> Thank you for
29 matches
Mail list logo