On Wed, Apr 11, 2018 at 2:30 PM, Wido den Hollander wrote:
>
>
> On 04/10/2018 09:45 PM, Gregory Farnum wrote:
>> On Tue, Apr 10, 2018 at 12:36 PM, Wido den Hollander wrote:
>>>
>>>
>>> On 04/10/2018 09:22 PM, Gregory Farnum wrote:
On Tue, Apr 10, 2018 at 6:32 AM Wido den Hollander >>>
Thanks for the suggestion but , unfortunately, having same number of OSD
did not solve the issue
Here is with 2 OSD per server, 3 servers - identical servers and osd
configuration
[root@osd01 ~]# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 4.02173 root default
-9
On 04/11/2018 07:48 PM, Steven Vacaroaia wrote:
Thanks for the suggestion but , unfortunately, having same number of
OSD did not solve the issue
Here is with 2 OSD per server, 3 servers - identical servers and osd
configuration
ceph osd pool ls detail
ceph osd crush rule dump
k
_
[root@osd01 ~]# ceph osd pool ls detail -f json-pretty
[
{
"pool_name": "rbd",
"flags": 1,
"flags_names": "hashpspool",
"type": 1,
"size": 2,
"min_size": 1,
"crush_rule": 0,
"object_hash": 2,
"pg_num": 128,
"pg_pla
Hi all,
We have a two-cluster-node (with a third "monitoring-only" node). Over
the last months, everything ran *perfectly* smooth. Today, I did an
Ubuntu "apt-get upgrade" on one of the two servers. Among others, the
ceph packages were upgraded from 12.2.1 to 12.2.2. A minor release
update, o
I think you have to update all osd's, mon's etc. I can remember running
into similar issue. You should be able to find more about this in
mailing list archive.
-Original Message-
From: Ranjan Ghosh [mailto:gh...@pw6.de]
Sent: woensdag 11 april 2018 16:02
To: ceph-users
Subject: [ceph
On Wed, Apr 11, 2018 at 2:43 AM, Mykola Golub wrote:
> On Tue, Apr 10, 2018 at 11:14:58PM -0400, Alex Gorbachev wrote:
>
>> So Josef fixed the one issue that enables e.g. lsblk and sysfs size to
>> reflect the correct siz on change. However, partptobe and parted
>> still do not detect the change,
Do you have a preliminary patch that we can test against?
On Wed, Apr 11, 2018 at 10:27 AM, Alex Gorbachev
wrote:
> On Wed, Apr 11, 2018 at 2:43 AM, Mykola Golub wrote:
>> On Tue, Apr 10, 2018 at 11:14:58PM -0400, Alex Gorbachev wrote:
>>
>>> So Josef fixed the one issue that enables e.g. lsblk
Hi all,
I use Pacemaker and the "Filesystem" Resource Agent to mount/unmount my
cephfs. Depending on timing, MDS may be reachable a few dozen seconds
after the mount command, but there is no report of the failure through
the exit code.
Examples using mount.fuse.ceph or ceph-fuse (no MDS running a
Thank you for your answer. Do you have any specifics on which thread
you're talking about? Would be very interested to read about a success
story, because I fear that if I update the other node that the whole
cluster comes down.
Am 11.04.2018 um 10:47 schrieb Marc Roos:
I think you have to u
> On Wed, Apr 11, 2018 at 10:27 AM, Alex Gorbachev
> wrote:
>> On Wed, Apr 11, 2018 at 2:43 AM, Mykola Golub
>> wrote:
>>> On Tue, Apr 10, 2018 at 11:14:58PM -0400, Alex Gorbachev wrote:
>>>
So Josef fixed the one issue that enables e.g. lsblk and sysfs size to
reflect the correct siz
Ah, nevermind, we've solved it. It was a firewall issue. The only thing
that's weird is that it became an issue immediately after an update.
Perhaps it has sth. to do with monitor nodes shifting around or
anything. Well, thanks again for your quick support, though. It's much
appreciated.
BR
Ok. How do I fix what's been broke? How do I "rebuild my index"?
Thanks
On Wed, Apr 11, 2018 at 1:49 AM, Robin H. Johnson
wrote:
> On Tue, Apr 10, 2018 at 10:06:57PM -0500, Robert Stanford wrote:
> > I used this command to purge my rgw data:
> >
> > rados purge default.rgw.buckets.data --
I'll give it a try locally and see if I can figure it out. Note that
this commit [1] also dropped the call to "bd_set_size" within
"nbd_size_update", which seems suspicious to me at initial glance.
[1]
https://github.com/torvalds/linux/commit/29eaadc0364943b6352e8994158febcb699c9f9b#diff-bc9273bc
ceph upgrades are usualy not a problem:
ceph have to be upgraded in the right order. normally when each service
is on its own machine this is not difficult.
but when you have mon, mgr, osd, mds, and klients on the same host you
have to do it a bit carefully..
i tend to have a terminal open wit
On 4/7/18 4:21 PM, Alfredo Deza wrote:
On Sat, Apr 7, 2018 at 11:59 AM, Gary Verhulp wrote:
I’m trying to create bluestore osds with separate --block.wal --block.db
devices on a write intensive SSD
I’ve split the SSD (/dev/sda) into two partditions sda1 and sda2 for db and
wal
I see
I've tested the patch on both 4.14.0 and 4.16.0 and it appears to
function correctly for me. parted can see the newly added free-space
after resizing the RBD image and our stress tests once again pass
successfully. Do you have any additional details on the issues you are
seeing?
On Wed, Apr 11, 20
Hello again,
I have reinstalled the cluster and noticed that, with 2 servers is working
as expectd, adding the 3rd one tanks perfermonce IRRESPECTIVE of which
server is the 3 rd one
I have tested it with only 1 OSD per server in order to eliminate any
balancing issues
This seems to indicate an is
On Wed, Apr 11, 2018 at 2:13 PM, Jason Dillaman wrote:
> I've tested the patch on both 4.14.0 and 4.16.0 and it appears to
> function correctly for me. parted can see the newly added free-space
> after resizing the RBD image and our stress tests once again pass
> successfully. Do you have any addi
On Tue, Apr 10, 2018 at 8:50 PM, Yan, Zheng wrote:
> On Wed, Apr 11, 2018 at 3:34 AM, Gregory Farnum wrote:
>> On Tue, Apr 10, 2018 at 5:54 AM, John Spray wrote:
>>> On Tue, Apr 10, 2018 at 1:44 PM, Yan, Zheng wrote:
Hello
To simplify snapshot handling in multiple active mds setu
Hello Ronny,
On Wed, Apr 11, 2018 at 10:25 AM, Ronny Aasen wrote:
> mds: restart mds's one at the time. you will notice the standby mds taking
> over for the mds that was restarted. do both.
No longer recommended. See:
http://docs.ceph.com/docs/master/cephfs/upgrading/#upgrading-the-mds-cluster
Hi,
For anybody who may be interested, here I share a process of locating the
reason for ceph cluster performance slow down in our environment.
Internally, we have a cluster with capacity 1.1PB, used 800TB, and raw user
data is about 500TB. Each day, 3TB' data is uploaded and 3TB oldest data i
On Wed, Apr 11, 2018 at 2:13 PM, Jason Dillaman wrote:
> I've tested the patch on both 4.14.0 and 4.16.0 and it appears to
> function correctly for me. parted can see the newly added free-space
> after resizing the RBD image and our stress tests once again pass
> successfully. Do you have any addi
After digging into our internal system stats, we find the new added's disk io
util is about two times than the old.
This is obviously and expected. Yours 8Tb drives weighted double against
4Tb and do *double* crush work in comparison.
k
___
cep
Yes, according to crush algorithms, large drives are given high weight, this is
expected. By default, crush gives no consideration of each drive's performance,
which may cause the performance distribution is not balanced. And the highest
io util osd may slow down the whole cluster.
On 04/12/2018 10:58 AM, ?? ? wrote:
Yes, according to crush algorithms, large drives are given high weight, this is
expected. By default, crush gives no consideration of each drive's performance,
which may cause the performance distribution is not balanced. And the highest
io util osd may slow
Currently, this can only be done by hand. Maybe we need some scripts to handle
this automatically.
I don't know if
https://github.com/ceph/ceph/tree/master/src/pybind/mgr/balancer can handle
this.
From: Konstantin Shalygin
Sent: Thursday, April 12, 2018
27 matches
Mail list logo