Re: [ceph-users] Ceph MDS WRN replayed op client.$id

2018-09-13 Thread Eugen Block
Hi Stefan, mds.mds1 [WRN] replayed op client.15327973:15585315,15585103 used ino 0x19918de but session next is 0x1873b8b Nothing of importance is logged in the mds (debug_mds_log": "1/5"). What does this warning message mean / indicate? we face these messages on a regular basis. The

Re: [ceph-users] Ceph MDS WRN replayed op client.$id

2018-09-13 Thread John Spray
On Wed, Sep 12, 2018 at 2:59 PM Stefan Kooman wrote: > > Hi, > > Once in a while, today a bit more often, the MDS is logging the > following: > > mds.mds1 [WRN] replayed op client.15327973:15585315,15585103 used ino > 0x19918de but session next is 0x1873b8b > > Nothing of importance is lo

[ceph-users] Not all pools are equal, but why

2018-09-13 Thread Stefan Kooman
Hi List, TL;DR: what application types are compatible with each other concerning Ceph Pools? I.e. is it safe to mix "RBD" pool with (some) native librados objects? RBD / RGW / Cephfs all have their own pools. Since luminous release there is this "application tag" to (somewhere in the future) pre

Re: [ceph-users] Not all pools are equal, but why

2018-09-13 Thread John Spray
On Thu, Sep 13, 2018 at 9:03 AM Stefan Kooman wrote: > > Hi List, > > TL;DR: what application types are compatible with each other concerning > Ceph Pools? > > I.e. is it safe to mix "RBD" pool with (some) native librados objects? > > RBD / RGW / Cephfs all have their own pools. Since luminous rel

[ceph-users] issues about module promethus

2018-09-13 Thread xiang . dai
Hi! I want to user promethus+grafana to monitor ceph, and I find below url: http: //docs.ceph.com/docs/master/mgr/prometheus/ Then i download ceph dashboard in grafana: https://grafana.com/dashboards/7056 It is so cool But some metrices do not work for ceph 13( Mimic ), like "ceph_monit

[ceph-users] issues about module promethus

2018-09-13 Thread xiang . dai
Hi! I want to user promethus+grafana to monitor ceph, and I find below url: http: //docs.ceph.com/docs/master/mgr/prometheus/ Then i download ceph dashboard in grafana: https://grafana.com/dashboards/7056 It is so cool But some metrices do not work for ceph 13( Mimic ), like "ceph_monit

Re: [ceph-users] Ceph MDS WRN replayed op client.$id

2018-09-13 Thread Stefan Kooman
Hi John, Quoting John Spray (jsp...@redhat.com): > On Wed, Sep 12, 2018 at 2:59 PM Stefan Kooman wrote: > > When replaying a journal (either on MDS startup or on a standby-replay > MDS), the replayed file creation operations are being checked for > consistency with the state of the replayed cli

Re: [ceph-users] Ceph MDS WRN replayed op client.$id

2018-09-13 Thread John Spray
On Thu, Sep 13, 2018 at 11:01 AM Stefan Kooman wrote: > > Hi John, > > Quoting John Spray (jsp...@redhat.com): > > > On Wed, Sep 12, 2018 at 2:59 PM Stefan Kooman wrote: > > > > When replaying a journal (either on MDS startup or on a standby-replay > > MDS), the replayed file creation operations

Re: [ceph-users] Rados performance inconsistencies, lower than expected performance

2018-09-13 Thread Menno Zonneveld
Update on the subject, warning, lengthy post but reproducible results and workaround to get performance back to expected level. One of the servers had a broken disk controller causing some performance issues on this one host, FIO showed about half performance on some disks compared to the other

Re: [ceph-users] Rados performance inconsistencies, lower than expected performance

2018-09-13 Thread Alwin Antreich
On Thu, Sep 13, 2018 at 02:17:20PM +0200, Menno Zonneveld wrote: > Update on the subject, warning, lengthy post but reproducible results and > workaround to get performance back to expected level. > > One of the servers had a broken disk controller causing some performance > issues on this one h

Re: [ceph-users] Rados performance inconsistencies, lower than expected performance

2018-09-13 Thread Menno Zonneveld
-Original message- > From:Alwin Antreich > Sent: Thursday 13th September 2018 14:41 > To: Menno Zonneveld > Cc: ceph-users ; Marc Roos > > Subject: Re: [ceph-users] Rados performance inconsistencies, lower than > expected performance > > > Am I doing something wrong? Did I run into so

Re: [ceph-users] RADOS async client memory usage explodes when reading several objects in sequence

2018-09-13 Thread Daniel Goldbach
I'm sure I'm not forgetting to free any buffers. I'm not even allocating any heap memory in the example above. On further investigation, the same issue *does* happen with the synchronous read operation API. I erroneously said that the issue doesn't happen with the synchronous API when what I meant

[ceph-users] Proxmox/ceph upgrade and addition of a new node/OSDs

2018-09-13 Thread Hervé Ballans
Dear list, I am currently in the process of upgrading Proxmox 4/Jewel to Proxmox5/Luminous. I also have a new node to add to my Proxmox cluster. What I plan to do is the following (from https://pve.proxmox.com/wiki/Ceph_Jewel_to_Luminous): * upgrade Jewel to Luminous * let the "ceph osd c

Re: [ceph-users] RADOS async client memory usage explodes when reading several objects in sequence

2018-09-13 Thread Gregory Farnum
On Thu, Sep 13, 2018 at 6:35 AM Daniel Goldbach wrote: > I'm sure I'm not forgetting to free any buffers. I'm not even allocating > any heap memory in the example above. > > On further investigation, the same issue *does* happen with the > synchronous read operation API. I erroneously said that t

Re: [ceph-users] RADOS async client memory usage explodes when reading several objects in sequence

2018-09-13 Thread Daniel Goldbach
Yes I understand that. If you look at the example, the data buffer is stack allocated and hence its memory is freed when the stack frame for readobj is destroyed. Additionally, no leak occurs if I comment out the rados_read_op_operate line. This is a problem with librados, not with my example. O

Re: [ceph-users] data corruption issue with "rbd export-diff/import-diff"

2018-09-13 Thread Patrick.Mclean
On 2018-09-12 19:49:16-07:00 Jason Dillaman wrote: On Wed, Sep 12, 2018 at 10:15 PM wrote: > > On 2018-09-12 17:35:16-07:00 Jason Dillaman wrote: > > > Any chance you know the LBA or byte offset of the corruption so I can > compare it against the log? &

Re: [ceph-users] data corruption issue with "rbd export-diff/import-diff"

2018-09-13 Thread Jason Dillaman
On Thu, Sep 13, 2018 at 1:54 PM wrote: > > On 2018-09-12 19:49:16-07:00 Jason Dillaman wrote: > > > On Wed, Sep 12, 2018 at 10:15 PM wrote: > > > > On 2018-09-12 17:35:16-07:00 Jason Dillaman wrote: > > > > > > Any chance you know the LBA or byte offset of

[ceph-users] Standby mgr stopped sending beacons after upgrade to 12.2.8

2018-09-13 Thread Christian Albrecht
Hi all, after upgrading from 12.2.7 to 12.2.8 the standby mgr instances in my cluster stopped sending beacons. The service starts and everything seems to work just fine, but after a period of time the mgr disappears. All of my three mgr daemons are running. [root@ceph01 ~]# ceph mgr dump {

Re: [ceph-users] Proxmox/ceph upgrade and addition of a new node/OSDs

2018-09-13 Thread mj
Hi Hervé, No answer from me, but just to say that I have exactly the same upgrade path ahead of me. :-) Please report here any tips, trics, or things you encountered doing the upgrades. It could potentially save us a lot of time. :-) Thanks! MJ On 09/13/2018 05:23 PM, Hervé Ballans wrote:

[ceph-users] Updating CRUSH Tunables to Jewel from Hammer

2018-09-13 Thread David Turner
I have a stage cluster with 4 HDDs and an SSD in each host. I have an EC profile that specifically chooses HDDs for placement. Also several Replica pools that write to either HDD or SSD. This has all worked well for a while. When I updated the Tunables to Jewel on the cluster, all of a sudden t

[ceph-users] lost osd while migrating EC pool to device-class crush rules

2018-09-13 Thread Graham Allan
I'm now following up to my earlier message regarding data migration from old to new hardware in our ceph cluster. As part of this we wanted to move to device-class-based crush rules. For the replicated pools the directions for this were straightforward; for our EC pool, it wasn't so clear, but

[ceph-users] can we drop support of centos/rhel 7.4?

2018-09-13 Thread kefu chai
hi ceph-{maintainers,users,developers}, recently, i ran into an issue[0] which popped up when we build Ceph on centos 7.5, but test it on centos 7.4. as we know, the gperftools-libs package provides the tcmalloc allocator shared library, but centos 7.4 and centos 7.5 ship different version of gper

[ceph-users] cephfs is growing up rapidly

2018-09-13 Thread Zhenshi Zhou
Hi, I have a ceph cluster of version 12.2.5 on centos7. I created 3 pools, 'rbd' for rbd storage, as well as 'cephfs_data' and 'cephfs_meta' for cephfs. Cephfs is used for backing up by rsync and volumes mounting by docker. The size of backup files is 3.5T. Besides, docker use less than 60G spac