[ceph-users] mds "Behing on trimming"

2016-03-21 Thread Dzianis Kahanovich
I have (second time) stuck mds warning: "Behind on trimming (63/30)". Looks working. What it mean and how to avoid it? And how to fix (exclude stop/migrate active mds)? Looks happened both time at night - probably on long backup/write operations (something like compressed local root backup to ceph

[ceph-users] Fwd: object unfound before backfill

2016-03-21 Thread lin zhou
Hi,guys my cluster face a network problem so it occur some error.after solve network problem. latency of some osds in one node is high,using ceph osd perf,which come to 3000+ so I delete this osd from cluster,keep osd data device. after recover and backfill,then I face the problem describe in ti

[ceph-users] object unfound before finish backfill, up set diff from acting set

2016-03-21 Thread lin zhou
Hi,guys my cluster face a network problem so it occur some error.after solve network problem. latency of some osds in one node is high,using ceph osd perf,which come to 3000+ so I delete this osd from cluster,keep osd data device. after recover and backfill,then I face the problem describe in ti

[ceph-users] Fresh install - all OSDs remain down and out

2016-03-21 Thread Markus Goldberg
Hi, i have upgraded my hardware and installed ceph totally new as described in http://docs.ceph.com/docs/master/rados/deployment/ The last job was creating the OSDs http://docs.ceph.com/docs/master/rados/deployment/ceph-deploy-osd/ I have used the create command and after that, the OSDs should b

[ceph-users] DSS 7000 for large scale object storage

2016-03-21 Thread Bastian Rosner
Hi, any chance that somebody here already got hands on Dell DSS 7000 machines? 4U chassis containing 90x 3.5" drives and 2x dual-socket server sleds (DSS7500). Sounds ideal for high capacity and density clusters, since each of the server-sleds would run 45 drives, which I believe is a suita

Re: [ceph-users] DSS 7000 for large scale object storage

2016-03-21 Thread David
Sounds like you’ll have a field day waiting for rebuild in case of a node failure or an upgrade of the crush map ;) David > 21 mars 2016 kl. 09:55 skrev Bastian Rosner : > > Hi, > > any chance that somebody here already got hands on Dell DSS 7000 machines? > > 4U chassis containing 90x 3.5"

Re: [ceph-users] cephfs infernalis (ceph version 9.2.1) - bonnie++

2016-03-21 Thread Yan, Zheng
On Mon, Mar 21, 2016 at 2:33 PM, Michael Hanscho wrote: > On 2016-03-21 05:07, Yan, Zheng wrote: >> On Sat, Mar 19, 2016 at 9:38 AM, Michael Hanscho wrote: >>> Hi! >>> >>> Trying to run bonnie++ on cephfs mounted via the kernel driver on a >>> centos 7.2.1511 machine resulted in: >>> >>> # bonnie

Re: [ceph-users] DSS 7000 for large scale object storage

2016-03-21 Thread Sean Redmond
I used a Unit a little like this ( https://www.sgi.com/products/storage/servers/mis_server.html) for a SATA pool in ceph - rebuilds after a failure of a node can be painful without a fair amount of testing & tuning. I have opted for more units with less disks for future builds using R730XD. On Mo

Re: [ceph-users] DONTNEED fadvise flag

2016-03-21 Thread Kenneth Waegeman
Thanks! As we are using the kernel client of EL7, does someone knows if that client supports it? On 16/03/16 20:29, Gregory Farnum wrote: On Wed, Mar 16, 2016 at 9:46 AM, Kenneth Waegeman wrote: Hi all, Quick question: Does cephFS pass the fadvise DONTNEED flag and take it into account? I wa

Re: [ceph-users] mds "Behing on trimming"

2016-03-21 Thread John Spray
On Mon, Mar 21, 2016 at 7:44 AM, Dzianis Kahanovich wrote: > I have (second time) stuck mds warning: "Behind on trimming (63/30)". Looks > working. What it mean and how to avoid it? And how to fix (exclude > stop/migrate > active mds)? The MDS has a metadata journal, whose length is measured in

Re: [ceph-users] Cannot remove rbd locks

2016-03-21 Thread Christoph Adomeit
Zhanks Jaseon, this worked ... On Fri, Mar 18, 2016 at 02:31:44PM -0400, Jason Dillaman wrote: > Try the following: > > # rbd lock remove vm-114-disk-1 "auto 140454012457856" client.71260575 > > -- > > Jason Dillaman > > > - Original Message - > > From: "Christoph Adomeit" > > To:

Re: [ceph-users] DSS 7000 for large scale object storage

2016-03-21 Thread Bastian Rosner
Yes, rebuild in case of a whole chassis failure is indeed an issue. That depends on how the failure domain looks like. I'm currently thinking of initially not running fully equipped nodes. Let's say four of these machines with 60x 6TB drives each, so only loaded 2/3. That's raw 1440TB distribu

Re: [ceph-users] mds "Behing on trimming"

2016-03-21 Thread Dzianis Kahanovich
PS Now I stop this mds, active migrated and warning removed. Cannot try more. Dzianis Kahanovich пишет: > John Spray пишет: > >>> Looks happened both time at night - probably on long backup/write operations >>> (something like compressed local root backup to cephfs). Also all local >>> mounts >>

Re: [ceph-users] Fresh install - all OSDs remain down and out

2016-03-21 Thread 施柏安
Hi Markus You should define the "osd device" and "host" then make ceph cluster work. Take the types in your map (osd, host, chasis.root) to design the crushmap according to your needed. Example: ​​ host node1 { id -1 alg straw hash 0 item osd.0 weight 1.00

Re: [ceph-users] DSS 7000 for large scale object storage

2016-03-21 Thread David
From my experience you’ll be better off planning exactly how many OSD’s and nodes you’re going to have and if possible equip them from the start. By just adding a new drive to the same pool ceph will start to rearrange data across the whole cluster which might lead to less client IO depending on

Re: [ceph-users] Does object map feature lock snapshots ?

2016-03-21 Thread Christoph Adomeit
vm and inside the vm run some io, i ran bonnie++ in a loop then go ahead and create first snapshot /usr/bin/rbd snap create rbd/vm-192-disk-1@initial.20160321-130439 export the snapshot (don't know if it is necessary) /usr/bin/rbd export --rbd-concurrent-management-ops 20 vm-192-disk-1@in

Re: [ceph-users] DONTNEED fadvise flag

2016-03-21 Thread Yan, Zheng
> On Mar 21, 2016, at 18:17, Kenneth Waegeman wrote: > > Thanks! As we are using the kernel client of EL7, does someone knows if that > client supports it? > fadvise DONTNEED is supported by kernel memory management subsystem. Fadvise DONTNEED works for all filesystems (including cephfs kern

[ceph-users] Ceph RBD client on OSD nodes - how about a Docker deployment?

2016-03-21 Thread Christian Sarrasin
Hi there, The docs have an ominous warning that one shouldn't run the RBD client (to mount block devices) on a machine which also serves OSDs [1] Due to budget constraints, this topology would be useful in our situation. Couple of q's: 1) Does the limitation also apply if the OSD daemon is

Re: [ceph-users] Ceph RBD client on OSD nodes - how about a Docker deployment?

2016-03-21 Thread Gregory Farnum
On Mon, Mar 21, 2016 at 11:45 AM, Christian Sarrasin wrote: > Hi there, > > The docs have an ominous warning that one shouldn't run the RBD client (to > mount block devices) on a machine which also serves OSDs [1] > > Due to budget constraints, this topology would be useful in our situation. > Cou

Re: [ceph-users] Ceph RBD client on OSD nodes - how about a Docker deployment?

2016-03-21 Thread Gregory Farnum
...wow. Sorry for the spam at this point. (How did you get some gmane address named after me to be in the recipients list?) On Mon, Mar 21, 2016 at 1:25 PM, Gregory Farnum wrote: > Heh, I failed to re-add the list the first time. Trying again, since > they can probably help more than me on this

[ceph-users] dependency of ceph_objectstore_tool in unhealthy ceph0.80.7 in ubuntu12.04

2016-03-21 Thread lin zhou
Hi, I want to using ceph_objectstore_tool to export a pg from an OSD which has been delete from cluster just as https://ceph.com/community/incomplete-pgs-oh-my/ do. my ceph version is 0.80.7,and ceph_objectstore_tool has a dependency of libgoogle-perftools0. But libgoogle-perftools4 has been i

[ceph-users] Optimations of cephfs clients on WAN: Looking for suggestions.

2016-03-21 Thread Goncalo Borges
Dear CephFS gurus... I would like your advise on how to improve performance without compromising reliability for CephFS clients deployed under a WAN. Currently, our infrastructure relies on: - ceph infernalis - a ceph object cluster, with all core infrastructure components sitting in the same d

Re: [ceph-users] Any suggestion to deal with slow request?

2016-03-21 Thread lin zhou
I face the same problem. my osd.7 occur slow request,and many pg has a stat of active+recovery_wait. I checked network and the device of osd.7,no errors. Have you solve your problem ? 2016-01-08 13:06 GMT+08:00 Christian Balzer : > > Hello, > > > On Fri, 8 Jan 2016 12:22:04 +0800 Jevon Qiao w