Hi Wyatt, You need to reduce the replication level on your existing pools to 1, or bring up another OSD. The default configuration specifies a replication level of 2, and the default crush rules want to place a replica on two distinct OSDs. With one OSD, CRUSH can't determine placement for the replica and so Ceph is reporting a degraded state.
Dino On Wed, May 1, 2013 at 11:45 AM, Wyatt Gorman <wyattgor...@wyattgorman.com>wrote: > Well, those points solved the issue of the redefined host and the > unidentified protocol. The > > > "HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery 21/42 > degraded (50.000%)" > > error is still an issue, though. Is this something simple like some hard > drive corruption that I can clean up with a fsck, or is this a ceph issue? > > > > On Wed, May 1, 2013 at 12:31 PM, Mike Dawson <mike.daw...@scholarstack.com > > wrote: > >> Wyatt, >> >> A few notes: >> >> - Yes, the second "host = ceph" under mon.a is redundant and should be >> deleted. >> >> - "auth client required = cephx [osd]" should be simply >> auth client required = cephx". >> >> - Looks like you only have one OSD. You need at least as many (and >> hopefully more) OSDs than highest replication level out of your pools. >> >> Mike >> >> >> On 5/1/2013 12:23 PM, Wyatt Gorman wrote: >> >>> Here is my ceph.conf. I just figured out that the second host = isn't >>> necessary, though it is like that on the 5-minute quick start guide... >>> (Perhaps I'll submit my couple of fixes that I've had to implement so >>> far). That fixes the "redefined host" issue, but none of the others. >>> >>> [global] >>> # For version 0.55 and beyond, you must explicitly enable or >>> # disable authentication with "auth" entries in [global]. >>> >>> auth cluster required = cephx >>> auth service required = cephx >>> auth client required = cephx [osd] >>> osd journal size = 1000 >>> >>> #The following assumes ext4 filesystem. >>> filestore xattr use omap = true >>> # For Bobtail (v 0.56) and subsequent versions, you may add >>> #settings for mkcephfs so that it will create and mount the file >>> #system on a particular OSD for you. Remove the comment `#` >>> #character for the following settings and replace the values in >>> #braces with appropriate values, or leave the following settings >>> #commented out to accept the default values. You must specify >>> #the --mkfs option with mkcephfs in order for the deployment >>> #script to utilize the following settings, and you must define >>> #the 'devs' option for each osd instance; see below. osd mkfs >>> #type = {fs-type} osd mkfs options {fs-type} = {mkfs options} # >>> #default for xfs is "-f" osd mount options {fs-type} = {mount >>> #options} # default mount option is "rw,noatime" >>> # For example, for ext4, the mount option might look like this: >>> >>> #osd mkfs options ext4 = user_xattr,rw,noatime >>> # Execute $ hostname to retrieve the name of your host, and >>> # replace {hostname} with the name of your host. For the >>> # monitor, replace {ip-address} with the IP address of your >>> # host. >>> [mon.a] >>> host = ceph >>> mon addr = 10.81.2.100:6789 <http://10.81.2.100:6789> [osd.0] >>> >>> host = ceph >>> >>> # For Bobtail (v 0.56) and subsequent versions, you may add >>> # settings for mkcephfs so that it will create and mount the >>> # file system on a particular OSD for you. Remove the comment >>> # `#` character for the following setting for each OSD and >>> # specify a path to the device if you use mkcephfs with the >>> # --mkfs option. >>> >>> #devs = {path-to-device} >>> [osd.1] >>> host = ceph >>> #devs = {path-to-device} >>> [mds.a] >>> host = ceph >>> >>> >>> On Wed, May 1, 2013 at 12:14 PM, Mike Dawson >>> <mike.daw...@scholarstack.com >>> <mailto:mike.dawson@**scholarstack.com<mike.daw...@scholarstack.com>>> >>> wrote: >>> >>> Wyatt, >>> >>> Please post your ceph.conf. >>> >>> - mike >>> >>> >>> On 5/1/2013 12:06 PM, Wyatt Gorman wrote: >>> >>> Hi everyone, >>> >>> I'm setting up a test ceph cluster and am having trouble getting >>> it >>> running (great for testing, huh?). I went through the >>> installation on >>> Debian squeeze, had to modify the mkcephfs script a bit because >>> it calls >>> monmaptool with too many paramaters in the $args variable (mine >>> had >>> "--add a [ip address]:[port] [osd1]" and I had to get rid of the >>> [osd1] >>> part for the monmaptool command to take it). Anyway, so I got it >>> installed, started the service, waiting a little while for it to >>> build >>> the fs, and ran "ceph health" and got (and am still getting >>> after a day >>> and a reboot) the following error: (note: I have also been >>> getting the >>> first line in various calls, unsure why it is complaining, I >>> followed >>> the instructions...) >>> >>> warning: line 34: 'host' in section 'mon.a' redefined >>> 2013-05-01 12:04:39.801102 b733b710 -1 WARNING: unknown auth >>> protocol >>> defined: [osd] >>> HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery >>> 21/42 >>> degraded (50.000%) >>> >>> Can anybody tell me the root of this issue, and how I can fix >>> it? Thank you! >>> >>> - Wyatt Gorman >>> >>> >>> ______________________________**___________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> <mailto:ceph-us...@lists.ceph.**com<ceph-users@lists.ceph.com> >>> > >>> >>> http://lists.ceph.com/__**listinfo.cgi/ceph-users-ceph._**_com<http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com> >>> >>> <http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**com<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> >>> > >>> >>> >>> > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- ______________________________ Dino Yancey 2GNT.com Admin
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com