Anyone? I've tried with 0.67-rc2 and it fails/hangs. My monitor directory is
now 200GB and I worried I'll lose this cluster with only one monitor.
Every time I go from 1 to 2 monitors, the monitors hang and stop responding. I
assume it looses quorum.
Here are the steps I take:
rm -rf /var/lib
Hi guys
I was wondering if there is any way to use rados_clone_range for
objects that reside in different placement groups.
Thanks
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Dear all,
I am having a issue with MON after reboot, I had originally 3 MONs but after
rebooting one of them, the quorum cannot be established. Digging through the
log of the problematic MON, I found the following messages:
2013-07-31 18:04:18.536331 7f4f41bfd700 0 -- 10.4.132.18:6804/0 >>
10
Hi,
I am having a issue with MON after reboot, I had originally 3 MONs but
after rebooting one of them, the quorum cannot be established. Digging
Did you happen to upgrade Ceph at some time where you haven't restarted
the daemon?
Which version of Ceph do you run on each server?
--
Jens Kr
Hi,
This happened to me twice actually. Once before I upgraded, using 0.61.4 (I
solved it by removing the MON and then recreating it) and this time after
upgrading all nodes to 0.61.7. The surprising thing to me is that the node is
also running OSD as well but after the upgrade and reboot, the
Hi,
This happened to me twice actually. Once before I upgraded, using
0.61.4 (I solved it by removing the MON and then recreating it) and
this time after upgrading all nodes to 0.61.7. The surprising thing
to me is that the node is also running OSD as well but after the
upgrade and reboot, the c
Hi,
Actually I am not sure if the upgrade (I am using Ubuntu 12.04) actually will
restart the daemon as a result of that I tried rebooting one of them first.
However, in a previous situation I had (while all of them were still running
the older 0.61.4). Rebooting the same MON, did not restore t
Hi,
However, in a previous situation I had (while all of them were still
running the older 0.61.4). Rebooting the same MON, did not restore
the quorum also, so I am not sure what may cause this but I supposed
it maybe some other reasons.
Actually it is the same reason.
When upgrading from a v
Hi List,
I want to deploy two ceph clusters on ubuntu 12.04 and centos 6.4 separately,
and test cephfs, krbd, and librbd.
Which Kernel and QEMU/libvirt version do you recommend? Any specific patches
which I should apply manually?
Thanks for your time!
Sorry, forgot to mention the ceph version I want to use.
I want to use the latest stable cuttlefish release, 0.61.7 currently.
-- Original --
From: "Da Chun";
Date: Wed, Jul 31, 2013 10:18 PM
To: "ceph-users";
Subject: [ceph-users] Which Kernel and QEMU/lib
Does anyone here have multiple clusters or segment their single cluster in
such a way as to try to maintain different SLAs for production vs
non-production services?
We have been toying with the idea of running separate clusters (on the same
hardware, but reserve a portion of the OSDs for the prod
On 07/31/2013 04:18 PM, Da Chun wrote:
Hi List,
I want to deploy two ceph clusters on ubuntu 12.04 and centos 6.4
separately, and test cephfs, krbd, and librbd.
Which Kernel and QEMU/libvirt version do you recommend? Any specific
patches which I should apply manually?
I'd recommend you use t
So after much testing, it appears there may be some residual data left on a
disk by a ceph installation? I can't think of another explanation.
I've done repeated installs using /dev/sdd, the first install worked, the rest
failed. Subsequent installs would appear to work, the OSD daemon would sta
I am running on Ubuntu 13.04.
There is something amiss with /etc/init.d/ceph on all of my ceph nodes.
I was upgrading to 0.61.7 from what I *thought* was 0.61.5 today when I
realized that "service ceph-all restart" wasn't actually doing anything. I
saw nothing in /var/log/ceph.log - it just kept
Greg,
You can check the currently running version (and much more) using the
admin socket:
http://ceph.com/docs/master/rados/operations/monitoring/#using-the-admin-socket
For me, this looks like:
# ceph --admin-daemon /var/run/ceph/ceph-mon.a.asok version
{"version":"0.61.7"}
# ceph --admin-
Greg,
IMO the most critical risks when running Ceph are bugs that affect
daemon stability and the upgrade process.
Due to the speed of releases in the Ceph project, I feel having separate
physical hardware is the safer way to go, especially in light of your
mention of an SLA for your product
On Wed, Jul 31, 2013 at 12:19 PM, Mike Dawson wrote:
> Due to the speed of releases in the Ceph project, I feel having separate
> physical hardware is the safer way to go, especially in light of your
> mention of an SLA for your production services.
>
Ah. I guess I should offer a little more back
Hi Greg,
I saw about the same thing on Ubuntu 13.04 as you did. I used
apt-get -y update
apt-get -y upgrade
On all my cluster nodes to upgrade from 0.61.5 to 0.61.7 and then
noticed
that some of my systems did not restart all the daemons. I tried:
stop ceph-all
start ceph-all
On those node
Blast and gadzooks. This is a bug then.
What's worse is that on three of my mon nodes have anything in
/var/run/ceph. The directory is empty! I can't believe I've basically
been running a busy ceph cluster for the last month.
I'll try what you suggested, thank you.
On Wed, Jul 31, 2013 at 3:
On 7/31/2013 3:34 PM, Greg Poirier wrote:
On Wed, Jul 31, 2013 at 12:19 PM, Mike Dawson mailto:mike.daw...@cloudapt.com>> wrote:
Due to the speed of releases in the Ceph project, I feel having
separate physical hardware is the safer way to go, especially in
light of your mention of
Hi,
What's worse is that on three of my mon nodes have anything in
/var/run/ceph. The directory is empty! I can't believe I've basically
been running a busy ceph cluster for the last month.
If you're mounting another filesystem on the mon directory mount point,
then check if that filesyste
Wild guess, but are you by chance using the ceph-run wrapper around
the daemons (enabled with docrun or --restart in the init script, if
memory serves)? I noticed similar strangeness (can't stop daemon,
can't check status) using ceph-run on a RHEL6-like distro a few months
back, with bobtail.
--
da
I just checked my cluster with 3 monitor nodes and 2 osd nodes, and
none of them
had sockets in /var/run/ceph. I verified
# ceph health
HEALTH_OK
So I rebooted one of my monitor nodes, and when it came back up, the
socket was there.
# ls -l /var/run/ceph/
total 0
srwxr-xr-x 1 root root 0 Ju
You do need to use the stop script, not service stop. If you use service
stop, Upstart will restart the service. It's ok for start and restart,
because that what you want anyway, but service stop is effectively a
restart.
I wouldn't recommend doing stop ceph-all and start ceph-all after an
u
Hi Greg, everyone,
I'm having trouble following this whole thread. Two questions:
- is there an upstart or sysvinit file in your /var/lib/ceph/mon/* dirs?
- are the daemons defined in [mon.xxx] sections in ceph.conf?
That will control whether it is sysvinit or upstart that should be doing
th
After I did what Eric Eastman, suggested, my mon and osd sockets showed up
in /var/run/ceph:
root@kvm-cs-sn-10i:/etc/ceph# ls /var/run/ceph/
ceph-osd.0.asok ceph-osd.1.asok ceph-osd.2.asok ceph-osd.3.asok
ceph-osd.4.asok ceph-osd.5.asok ceph-osd.6.asok ceph-osd.7.asok
However, while the os
I'm having trouble following this whole thread. Two questions:
>
> - is there an upstart or sysvinit file in your /var/lib/ceph/mon/* dirs?
>
upstart. My first mistake was that I was trying to use both upstart and
sysv, depending on what directions I was following from the list, IRC, or
site d
For CentOS 6.4, we have custom qemu packages available at
http://ceph.com/packages/ceph-extras/rpm/centos6 which will provide RBD
support.
You will need to install a newer kernel than the one which ships by default
(2.6.32) to use the cephfs or krbd drivers. Any version above 3.x should be
sufficie
Hi,
I just added a second node to my ceph test platform. The first node has
a mon and three osd's, the second node only has three osd's. Adding the
osd's was pretty painless, and ceph distributed the data from the first
node evenly over both nodes so everything seems to be fine. The monitor
also t
Thanks! Neil and Wido.
Neil, what about the livirt version on CentOS 6.4? Just use the official
release?
-- Original --
From: "Neil Levine";
Date: Thu, Aug 1, 2013 05:53 AM
To: "Da Chun";
Cc: "ceph-users";
Subject: Re: [ceph-users] Which Kernel and QEMU/l
Yes, default version should work.
Neil
On Wed, Jul 31, 2013 at 7:11 PM, Da Chun wrote:
> Thanks! Neil and Wido.
>
> Neil, what about the livirt version on CentOS 6.4? Just use the official
> release?
>
> -- Original --
> *From: * "Neil Levine";
> *Date: * Thu, A
Logging might well help.
http://ceph.com/docs/master/rados/troubleshooting/log-and-debug/
On 07/31/2013 03:51 PM, Erik Logtenberg wrote:
Hi,
I just added a second node to my ceph test platform. The first node has
a mon and three osd's, the second node only has three osd's. Adding the
osd's w
32 matches
Mail list logo