I have (second time) stuck mds warning: "Behind on trimming (63/30)". Looks
working. What it mean and how to avoid it? And how to fix (exclude stop/migrate
active mds)?
Looks happened both time at night - probably on long backup/write operations
(something like compressed local root backup to ceph
Hi,guys
my cluster face a network problem so it occur some error.after solve
network problem.
latency of some osds in one node is high,using ceph osd perf,which come to
3000+
so I delete this osd from cluster,keep osd data device.
after recover and backfill,then I face the problem describe in ti
Hi,guys
my cluster face a network problem so it occur some error.after solve
network problem.
latency of some osds in one node is high,using ceph osd perf,which come to 3000+
so I delete this osd from cluster,keep osd data device.
after recover and backfill,then I face the problem describe in
ti
Hi,
i have upgraded my hardware and installed ceph totally new as described
in http://docs.ceph.com/docs/master/rados/deployment/
The last job was creating the OSDs
http://docs.ceph.com/docs/master/rados/deployment/ceph-deploy-osd/
I have used the create command and after that, the OSDs should b
Hi,
any chance that somebody here already got hands on Dell DSS 7000
machines?
4U chassis containing 90x 3.5" drives and 2x dual-socket server sleds
(DSS7500). Sounds ideal for high capacity and density clusters, since
each of the server-sleds would run 45 drives, which I believe is a
suita
Sounds like you’ll have a field day waiting for rebuild in case of a node
failure or an upgrade of the crush map ;)
David
> 21 mars 2016 kl. 09:55 skrev Bastian Rosner :
>
> Hi,
>
> any chance that somebody here already got hands on Dell DSS 7000 machines?
>
> 4U chassis containing 90x 3.5"
On Mon, Mar 21, 2016 at 2:33 PM, Michael Hanscho wrote:
> On 2016-03-21 05:07, Yan, Zheng wrote:
>> On Sat, Mar 19, 2016 at 9:38 AM, Michael Hanscho wrote:
>>> Hi!
>>>
>>> Trying to run bonnie++ on cephfs mounted via the kernel driver on a
>>> centos 7.2.1511 machine resulted in:
>>>
>>> # bonnie
I used a Unit a little like this (
https://www.sgi.com/products/storage/servers/mis_server.html) for a SATA
pool in ceph - rebuilds after a failure of a node can be painful without a
fair amount of testing & tuning.
I have opted for more units with less disks for future builds using R730XD.
On Mo
Thanks! As we are using the kernel client of EL7, does someone knows if
that client supports it?
On 16/03/16 20:29, Gregory Farnum wrote:
On Wed, Mar 16, 2016 at 9:46 AM, Kenneth Waegeman
wrote:
Hi all,
Quick question: Does cephFS pass the fadvise DONTNEED flag and take it into
account?
I wa
On Mon, Mar 21, 2016 at 7:44 AM, Dzianis Kahanovich
wrote:
> I have (second time) stuck mds warning: "Behind on trimming (63/30)". Looks
> working. What it mean and how to avoid it? And how to fix (exclude
> stop/migrate
> active mds)?
The MDS has a metadata journal, whose length is measured in
Zhanks Jaseon,
this worked ...
On Fri, Mar 18, 2016 at 02:31:44PM -0400, Jason Dillaman wrote:
> Try the following:
>
> # rbd lock remove vm-114-disk-1 "auto 140454012457856" client.71260575
>
> --
>
> Jason Dillaman
>
>
> - Original Message -
> > From: "Christoph Adomeit"
> > To:
Yes, rebuild in case of a whole chassis failure is indeed an issue. That
depends on how the failure domain looks like.
I'm currently thinking of initially not running fully equipped nodes.
Let's say four of these machines with 60x 6TB drives each, so only
loaded 2/3.
That's raw 1440TB distribu
PS Now I stop this mds, active migrated and warning removed. Cannot try more.
Dzianis Kahanovich пишет:
> John Spray пишет:
>
>>> Looks happened both time at night - probably on long backup/write operations
>>> (something like compressed local root backup to cephfs). Also all local
>>> mounts
>>
Hi Markus
You should define the "osd device" and "host" then make ceph cluster work.
Take the types in your map (osd, host, chasis.root) to design the
crushmap according to your needed.
Example:
host node1 {
id -1
alg straw
hash 0
item osd.0 weight 1.00
From my experience you’ll be better off planning exactly how many OSD’s and
nodes you’re going to have and if possible equip them from the start.
By just adding a new drive to the same pool ceph will start to rearrange data
across the whole cluster which might lead to less client IO depending on
vm and inside the vm run some io, i ran bonnie++ in a loop
then go ahead and create first snapshot
/usr/bin/rbd snap create rbd/vm-192-disk-1@initial.20160321-130439
export the snapshot (don't know if it is necessary)
/usr/bin/rbd export --rbd-concurrent-management-ops 20
vm-192-disk-1@in
> On Mar 21, 2016, at 18:17, Kenneth Waegeman wrote:
>
> Thanks! As we are using the kernel client of EL7, does someone knows if that
> client supports it?
>
fadvise DONTNEED is supported by kernel memory management subsystem. Fadvise
DONTNEED works for all filesystems (including cephfs kern
Hi there,
The docs have an ominous warning that one shouldn't run the RBD client
(to mount block devices) on a machine which also serves OSDs [1]
Due to budget constraints, this topology would be useful in our
situation. Couple of q's:
1) Does the limitation also apply if the OSD daemon is
On Mon, Mar 21, 2016 at 11:45 AM, Christian Sarrasin
wrote:
> Hi there,
>
> The docs have an ominous warning that one shouldn't run the RBD client (to
> mount block devices) on a machine which also serves OSDs [1]
>
> Due to budget constraints, this topology would be useful in our situation.
> Cou
...wow. Sorry for the spam at this point.
(How did you get some gmane address named after me to be in the
recipients list?)
On Mon, Mar 21, 2016 at 1:25 PM, Gregory Farnum wrote:
> Heh, I failed to re-add the list the first time. Trying again, since
> they can probably help more than me on this
Hi,
I want to using ceph_objectstore_tool to export a pg from an OSD which
has been delete from cluster just as
https://ceph.com/community/incomplete-pgs-oh-my/ do.
my ceph version is 0.80.7,and ceph_objectstore_tool has a dependency
of libgoogle-perftools0.
But libgoogle-perftools4 has been i
Dear CephFS gurus...
I would like your advise on how to improve performance without compromising
reliability for CephFS clients deployed under a WAN.
Currently, our infrastructure relies on:
- ceph infernalis
- a ceph object cluster, with all core infrastructure components sitting in the
same d
I face the same problem.
my osd.7 occur slow request,and many pg has a stat of active+recovery_wait.
I checked network and the device of osd.7,no errors.
Have you solve your problem ?
2016-01-08 13:06 GMT+08:00 Christian Balzer :
>
> Hello,
>
>
> On Fri, 8 Jan 2016 12:22:04 +0800 Jevon Qiao w
23 matches
Mail list logo