I am afraid, I hit the same bug. Giant worked fine, but after upgrading to
hammer (0.94.1) and putting some load on it, the MDSs eventually crashed and
now I am stuck in clientreplay most of the time. I am also using the cephfs
kernel client (3.18.y). As I didn't find a corresponding tracker ent
On Tue, May 19, 2015 at 4:31 PM, Markus Blank-Burian wrote:
> I am afraid, I hit the same bug. Giant worked fine, but after upgrading to
> hammer (0.94.1) and putting some load on it, the MDSs eventually crashed and
> now I am stuck in clientreplay most of the time. I am also using the cephfs
> ke
Here are some logs and the infos from the mdsc files. But I am afraid that
there might not be much info in the logs, since I had a very low log level.
Look for example at 2015-05-18T21:28:33+02:00. The mdsc files are concatenated
from all of the clients.
> Date: Tue, 19 May 2015 16:45:12 +0
On 2015年05月19日 17:07, Markus Blank-Burian wrote:
Here are some logs and the infos from the mdsc files. But I am afraid
that there might not be much info in the logs, since I had a very low
log level. Look for example at2015-05-18T21:28:33+02:00. The mdsc files
are concatenated from all of the cli
could you try the attached patch
On Tue, May 19, 2015 at 5:10 PM, Markus Blank-Burian wrote:
> Forgot the attachments. Besides, is there any way to get the cluster
> running again without restarting all client nodes?
>
> On Tue, May 19, 2015 at 10:45 AM, Yan, Zheng wrote:
>> On Tue, May 19, 2015
Hi!
The QEMU Venom vulnerability (http://venom.crowdstrike.com/) got my
attention and I would
like to know what are you people doing in order to have the latest
patched QEMU version
working with Ceph RBD?
In my case I am using the qemu-img and qemu-kvm packages provided by
Ceph (http://ceph.
Thanks for the patch! Testing might take up to a week, since I have to
reboot all the client nodes in the computing cluster.
On Tue, May 19, 2015 at 12:27 PM, Yan, Zheng wrote:
> could you try the attached patch
>
> On Tue, May 19, 2015 at 5:10 PM, Markus Blank-Burian
> wrote:
>> Forgot the att
Last night our Hammer cluster suffered a series of OSD crashes on all cluster
nodes.
We were running Hammer (0.94.1-98-g7df3eb5, built because we had a major
problem a week ago which we suspected to be related to bugs we found in the
tracker, that were not yet in 0.94.1). In the meantime we down
I actually managed to reboot everything today and it ran smoothly for
the last few minutes. MDS failover also worked without problems. If
anything bad happens in the next days, I will let you know.
Markus
On Tue, May 19, 2015 at 1:12 PM, Markus Blank-Burian wrote:
> Thanks for the patch! Testing
All,
I have looked at the various guides in ceph.com related to building and
deploying Calamari, and they will build the RPMs based on CentOS 6 or
Red Hat 7, but not CentOS 7.
I have no problem with the OS of the Calamari server, as I am thinking
in creating a VM for this, and Ubuntu or Cent
Hi!
Seeking performance improvement in our cluster (Firefly 0.80.7 on Wheezy, 5
nodes, 58 osds), I wrote
a small python script, that walks through ceph nodes and issue 'perf dump'
command on osd admin
sockets. It extracts *_latency tuples, calculate min/max/avg, compare osd perf
metrics with mi
You appear to be using pool snapshots with radosgw, I suspect that's what is
causing the issue. Can you post a longer log? Preferably with
debug osd = 20
debug filestore = 20
debug ms = 1
from startup to crash on an osd?
-Sam
- Original Message -
From: "Daniel Schneller"
To: ceph-use
Hello,
this question was brought many times before, and also solved in a
various ways - snap trimmer, scheduler` priorities and persistent fix
(for a ReplicatedPG issue), but it seems that the current Ceph
versions may suffer as well during the rollback operations on large
images and on large scal
I have 3 GW servers, but they are defined like this:
[client.radosgw.ceph-gw1]
rgw_ops_log_data_backlog = 4096
rgw_enable_ops_log = true
keyring = /etc/ceph/ceph.client.radosgw.keyring
rgw_print_continue = true
rgw_ops_log_rados = true
host = ceph-gw1
rgw_frontends = civetweb port=80
rgw_
Hi,
I am seeing write performance hit with small files (60K) using radosgw. The
radosgw is configured to run with 600 threads. Here is the write speed I
get with file sizes of 60K
# sudo ceph -s
cluster e445e46e-4d84-4606-9923-16fff64446dc
health HEALTH_OK
monmap e1: 1 mons at {osd
Hi list,
I was struggeling quiet a while with the problem that on my cephfs data pool
some PG’s stays inconsistent and could not be repaired. The message in OSD’s
log was like
>> repair 11.23a 57b4363a/2015b67.06e1/head//11 on disk data digest
>> 0x325d0322 != 0xe8c0243
and then the
I am trying to build the packages manually and I was wondering
is the flag --enable-rbd enough to have full Ceph functionality?
Does anybody know what else flags should I include in order to have the
same functionality as the original CentOS package plus the RBD support?
Regards,
George
On T
Been doing some more digging. I'm getting messages in the OSD logs like
these, don't know if these are normal or a clue to something not right
2015-05-19 18:36:27.664698 7f58b91dd700 0 log_channel(cluster) log [WRN] :
slow request 30.346117 seconds old, received at 2015-05-19 18:35:57.318208:
osd
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA256
You should be able to get the SRPM, extract the SPEC file and use that
to build a new package. You should be able to tweak all the compile
options as well. I'm still really new to building/rebuilding RPMs but
I've been able to do this for a couple of
You can also just fetch the rhev SRPMs and build those. They have rbd
enabled already.
On May 19, 2015 12:31 PM, "Robert LeBlanc" wrote:
> -BEGIN PGP SIGNED MESSAGE-
> Hash: SHA256
>
> You should be able to get the SRPM, extract the SPEC file and use that
> to build a new package. You sh
Hello,
So here are the steps I performed and where I sit now.
Step 1) Using 'ceph-objectstore-tool list' to create a list of all PGs not
associated with the 3 pools (rbd, data, metadata) that are actually in use
on this cluster.
Step 2) I then did a 'ceph-objectstore-tool remove' of those PGs
T
If 2.14 is part of a non-existent pool, you should be able to rename it out of
current/ in the osd directory to prevent the osd from seeing it on startup.
-Sam
- Original Message -
From: "Berant Lemmenes"
To: "Samuel Just"
Cc: ceph-users@lists.ceph.com
Sent: Tuesday, May 19, 2015 12:58:
Erik,
are you talking about the ones here :
http://ftp.redhat.com/redhat/linux/enterprise/6Server/en/RHEV/SRPMS/ ???
From what I see the version is rather "small" 0.12.1.2-2.448
How one can verify that it has been patched against venom
vulnerability?
Additionally I only see the qemu-kvm pa
Sorry, I made the assumption you were on 7. If you're on 6 then I defer to
someone else ;)
If you're on 7, go here.
http://ftp.redhat.com/pub/redhat/linux/enterprise/7Server/en/RHEV/SRPMS/
On May 19, 2015 2:47 PM, "Georgios Dimitrakakis"
wrote:
> Erik,
>
> are you talking about the ones here :
Erik,
thanks for the feedback. I am still on 6 so if someone else has a
proposal please come forward...
Best,
George
Sorry, I made the assumption you were on 7. If youre on 6 then I
defer
to someone else ;)
If youre on 7, go here.
http://ftp.redhat.com/pub/redhat/linux/enterprise/7Serve
Sam,
It is for a valid pool, however the up and acting sets for 2.14 both show
OSDs 8 & 7. I'll take a look at 7 & 8 and see if they are good.
If so, it seems like it being present on osd.3 could be an artifact from
previous topologies and I could mv it off old.3
Thanks very much for the assist
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA256
I've downloaded the new tarball, placed it in rpmbuild/SOURCES then
with the extracted spec file in rpmbuild/SPEC, I update it to the new
version and then rpmbuild -ba program.spec. If you install the SRPM
then it will install the RH patches that hav
Hi,
I'm hoping for advice on whether Ceph could be used in an atypical use
case. Specifically, I have about ~20TB of files that need replicated to 2
different sites. Each site has its own internal gigabit ethernet network.
However, the connection between the sites is only ~320kbits. I'm trying to
28 matches
Mail list logo