hi all,
I have some very basic questions about pools in ceph.
According to ceph documentation, as we deploy a ceph cluster with radosgw
instance over it, ceph creates pool by default to store the data or the
deployer can also create pools according to the requirement.
Now, my question is:
1. w
[root@chefwks01 ~]# ceph --cluster us-west01 osd crush dump
{ "devices": [
{ "id": 0,
"name": "osd.0"},
{ "id": 1,
"name": "osd.1"},
{ "id": 2,
"name": "osd.2"},
{ "id": 3,
"name": "osd.3"},
{ "id": 4,
That is CEPH_FEATURE_CRUSH_V2. Can you attach teh output of
ceph osd crush dump
Thanks!
sage
On Thu, 10 Jul 2014, Joshua McClintock wrote:
> Yes, I change some of the mount options on my osds (xfs mount options), but
> I think this may be the answer from dmesg, sorta looks like a version
> m
Hi,
I try to create a qemu image, but got failed.
ceph@ceph:~/my-cluster$ qemu-img create -f rbd rbd:rbd/qemu 2G
Formatting 'rbd:rbd/qemu', fmt=rbd size=2147483648 cluster_size=0
qemu-img: error connecting
qemu-img: rbd:rbd/qemu: error while creating rbd: Input/output error
Can you tell what's
Hi Craig,
On Thu, Jul 10, 2014 at 03:09:51PM -0700, Craig Lewis wrote:
> I fixed this issue by reformatting all of the OSDs. I changed the mkfs
> options from
>
> [osd]
> osd mkfs type = xfs
> osd mkfs options xfs = -l size=1024m -n size=64k -i size=2048 -s size=4096
>
> to
> [osd]
> osd
Yes, I change some of the mount options on my osds (xfs mount options), but
I think this may be the answer from dmesg, sorta looks like a version
mismatch:
libceph: loaded (mon/osd proto 15/24)
ceph: loaded (mds proto 32)
libceph: mon0 192.168.0.14:6789 feature set mismatch, my 4a042aca <
server
Have you made any other changes after the upgrade? (Like adjusting
tunables, or creating EC pools?)
See if there is anything in 'dmesg' output.
sage
On Thu, 10 Jul 2014, Joshua McClintock wrote:
> I upgraded my cluster to .80.1-2 (CentOS). My mount command just freezes
> and outputs an error
I upgraded my cluster to .80.1-2 (CentOS). My mount command just freezes
and outputs an error:
mount.ceph 192.168.0.14,192.168.0.15,192.168.0.16:/ /us-west01 -o
name=chefwks01,secret=`ceph-authtool -p -n client.admin
/etc/ceph/us-west01.client.admin.keyring`
mount error 5 = Input/output error
Hi,
I prefer to use bCache or other likely local write back cache on SSD since
it is only related to local HDDs. It think, it will reduce the risk of
error on cache flushing comparing to CEPH Cache Tiering which still using
clustering network on flushing. After the data has been written to the
cac
Hi,
I prefer to use bcache or other likely local write back cache on SSD since
it is only related to local HDDs. It think, it will reduce the risk of
error on cache flushing comparing to CEPH Cache Tiering which still using
clustering network on flushing. After the data has been written to the
cac
It could be an indication of a problem on osd 5, but the timing is
worrying. Can you attach your ceph.conf? Have there been any osds
going down, new osds added, anything to cause recovery? Anything in
dmesg to indicate an fs problem? Have you recently changed any
settings?
-Sam
On Thu, Jul 10,
Hi,
I've just upgraded a ceph cluster from Ubuntu 12.04 with 0.73.1 to Ubuntu
14.04 with 0.80.1.
I've noticed that the log rotation doesn't appear to work correctly.
The OSD's are just not logging to the current ceph-osd-X.log file.
If I restart the OSD's, they start logging, but then overnight,
I fixed this issue by reformatting all of the OSDs. I changed the mkfs
options from
[osd]
osd mkfs type = xfs
osd mkfs options xfs = -l size=1024m -n size=64k -i size=2048 -s size=4096
to
[osd]
osd mkfs type = xfs
osd mkfs options xfs = -s size=4096
(I have a mix of 512 and 4k sector dr
Greetings,
Just a follow up on my original issue. =ceph pg repair ...= fixed the
problem. However, today I got another inconsistent pg. It's interesting to
me that this second error is in the same rbd image and appears to be
"close" to the previously inconsistent pg. (Even more fun, osd.5 was the
Thanks - so it appears that the advantage of the 3rd replica (relative to 2
replicas) has to do much more with recovering from two concurrent OSD failures
than with inconsistencies found during deep scrub - would you agree?
Re: repair - do you mean the "repair" process during deep scrub - if ye
Repair I think will tend to choose the copy with the lowest osd number
which is not obviously corrupted. Even with three replicas, it does
not do any kind of voting at this time.
-Sam
On Thu, Jul 10, 2014 at 10:39 AM, Chahal, Sudip wrote:
> I've a basic related question re: Firefly operation - w
I've a basic related question re: Firefly operation - would appreciate any
insights:
With three replicas, if checksum inconsistencies across replicas are found
during deep-scrub then:
a. does the majority win or is the primary always the winner and used
to overwrite the secondaries
Can you attach your ceph.conf for your osds?
-Sam
On Thu, Jul 10, 2014 at 8:01 AM, Christian Eichelmann
wrote:
> I can also confirm that after upgrading to firefly both of our clusters
> (test and live) were going from 0 scrub errors each for about 6 Month to
> about 9-12 per week...
> This also
On Thu, Jul 10, 2014 at 9:04 AM, Joe Hewitt wrote:
> Hi there
> Recently I got a problem triggered by rebooting ceph nodes, which eventually
> wound up by rebuilding from ground up. A too-long-don't-read question here
> is: is there suggested best practices for online/offline ceph node?
>
> Follow
Hi there
Recently I got a problem triggered by rebooting ceph nodes, which eventually
wound up by rebuilding from ground up. A too-long-don't-read question here
is: is there suggested best practices for online/offline ceph node?
Following the official ceph doc, I set up a 4 node ceph (firefly) cl
That makes two of us...
G.
On Thu, 10 Jul 2014 17:12:08 +0200 (CEST), Alexandre DERUMIER wrote:
Ok, sorry, we have finally receive the login a bit late.
Sorry again to have spam the mailing list
- Mail original -
De: "Alexandre DERUMIER"
À: "ceph-users"
Envoyé: Jeudi 10 Juillet 2014
Hi,
I've just upgraded a ceph cluster from Ubuntu 12.04 with 0.72.1 to Ubuntu
14.04 with 0.80.1.
I've noticed that the log rotation doesn't appear to work correctly.
The OSD's are just not logging to the current ceph-osd-X.log file.
If I restart the OSD's or run "service ceph-osd reload id=X", th
On Thursday, July 10, 2014, Erik Logtenberg wrote:
>
> > Yeah, Ceph will never voluntarily reduce the redundancy. I believe
> > splitting the "degraded" state into separate "wrongly placed" and
> > "degraded" (reduced redundancy) states is currently on the menu for
> > the Giant release, but it's
Ok, sorry, we have finally receive the login a bit late.
Sorry again to have spam the mailing list
- Mail original -
De: "Alexandre DERUMIER"
À: "ceph-users"
Envoyé: Jeudi 10 Juillet 2014 16:55:22
Objet: [ceph-users] inktank-mellanox webinar access ?
Hi,
sorry to spam the mailin
The same here...Neither do I or my colleagues
G.
On Thu, 10 Jul 2014 16:55:22 +0200 (CEST), Alexandre DERUMIER wrote:
Hi,
sorry to spam the mailing list,
but they are a inktank mellanox webinar in 10minutes,
and I don't have receive access since I have been registered
yesterday (same for my
Hello to all,
I was wondering is it possible to place different pools on different
OSDs, but using only two physical servers?
I was thinking about this: http://tinypic.com/r/30tgt8l/8
I would like to use osd.0 and osd.1 for Cinder/RBD pool, and osd.2 and
osd.3 for Nova instances. I was follo
I can also confirm that after upgrading to firefly both of our clusters (test
and live) were going from 0 scrub errors each for about 6 Month to about 9-12
per week...
This also makes me kind of nervous, since as far as I know everything "ceph pg
repair" does, is to copy the primary object to al
Hi,
sorry to spam the mailing list,
but they are a inktank mellanox webinar in 10minutes,
and I don't have receive access since I have been registered yesterday (same
for my co-worker).
and the webinar mellanox contact email (conta...@mellanox.com), does not
exist
Maybe somebody from In
Hi,
I am getting some troubles with the ceph mon stability.
Every couple of days mons die. I only see this error in the logs:
2014-07-08 14:24:53.056805 7f713bb5b700 -1 mon.cephmon02@1(peon) e2 ***
Got Signal Interrupt ***
2014-07-08 14:24:53.061795 7f713bb5b700 1 mon.cephmon02@1(peon) e2 s
OK - I found this information
(http://comments.gmane.org/gmane.comp.file-systems.ceph.user/4992):
When creating a bucket users can specify which placement target they
want to use with that specific bucket (by using the S3 create bucket
location constraints field, format is
location_constraints='[
Hi,
we try to set up a Webcluster, may somebody give a hint.
3 Webserver with Typo3
The Typo3-Cache on a central Storage with ceph.
May this be usefull within a vmware-cluster?
I need something different than a central NFS-Store. This is to slow.
Regards
Marc
And actually just to follow-up, it does seem like there are some additional
smarts beyond just using the primary to overwrite the secondaries... Since
I captured md5 sums before and after the repair, I can say that in this
particular instance, the secondary copy was used to overwrite the primary.
I can also say that after a recent upgrade to Firefly, I have experienced
massive uptick in scrub errors. The cluster was on cuttlefish for about a
year, and had maybe one or two scrub errors. After upgrading to Firefly,
we've probably seen 3 to 4 dozen in the last month or so (was getting 2-3 a
Hi All,
There is a new bug-fix release of ceph-deploy, the easy deployment tool
for Ceph.
The full list of fixes for this release can be found in the changelog:
http://ceph.com/ceph-deploy/docs/changelog.html#id1
Make sure you update!
___
ceph-users
> Yeah, Ceph will never voluntarily reduce the redundancy. I believe
> splitting the "degraded" state into separate "wrongly placed" and
> "degraded" (reduced redundancy) states is currently on the menu for
> the Giant release, but it's not been done yet.
That would greatly improve the accuracy o
Hi,
RHEL7 repository works just as well. CentOS 7 is effectively a copy of
RHEL7 anyway. Packages for CentOS 7 wouldn't actually be any different.
Erik.
On 07/10/2014 06:14 AM, Alexandre DERUMIER wrote:
> Hi,
>
> I would like to known if a centos7 respository will be available soon ?
>
> Or ca
On 07/10/2014 03:24 AM, Xabier Elkano wrote:
El 10/07/14 09:18, Christian Balzer escribió:
On Thu, 10 Jul 2014 08:57:56 +0200 Xabier Elkano wrote:
El 09/07/14 16:53, Christian Balzer escribió:
On Wed, 09 Jul 2014 07:07:50 -0500 Mark Nelson wrote:
On 07/09/2014 06:52 AM, Xabier Elkano wrote:
On 07/09/2014 09:53 AM, Christian Balzer wrote:
On Wed, 09 Jul 2014 07:07:50 -0500 Mark Nelson wrote:
On 07/09/2014 06:52 AM, Xabier Elkano wrote:
El 09/07/14 13:10, Mark Nelson escribió:
On 07/09/2014 05:57 AM, Xabier Elkano wrote:
Hi,
I was doing some tests in my cluster with fio tool,
Hi,
Great.
All my OSD restart :
osdmap e438044: 36 osds: 36 up, 36 in
All PG page are active and some in recovery :
1604040/49575206 objects degraded (3.236%)
1780 active+clean
17 active+degraded+remapped+backfilling
61 active+degraded+remapped+wait_backfill
11 active+clean+scrubbing+deep
El 10/07/14 09:18, Christian Balzer escribió:
> On Thu, 10 Jul 2014 08:57:56 +0200 Xabier Elkano wrote:
>
>> El 09/07/14 16:53, Christian Balzer escribió:
>>> On Wed, 09 Jul 2014 07:07:50 -0500 Mark Nelson wrote:
>>>
On 07/09/2014 06:52 AM, Xabier Elkano wrote:
> El 09/07/14 13:10, Mark Ne
On Thu, 10 Jul 2014 08:57:56 +0200 Xabier Elkano wrote:
> El 09/07/14 16:53, Christian Balzer escribió:
> > On Wed, 09 Jul 2014 07:07:50 -0500 Mark Nelson wrote:
> >
> >> On 07/09/2014 06:52 AM, Xabier Elkano wrote:
> >>> El 09/07/14 13:10, Mark Nelson escribió:
> On 07/09/2014 05:57 AM, Xabi
41 matches
Mail list logo