Hi Wido.

Thanks again.

I will rebuild the cluster with bigger disk.

Again thanks for your help.


2015-07-13 14:15 GMT+02:00 Wido den Hollander <w...@42on.com>:

>
>
> On 13-07-15 14:07, alberto ayllon wrote:
> > On 13-07-15 13:12, alberto ayllon wrote:
> >> Maybe this can help to get the origin of the problem.
> >>
> >> If I run  ceph pg dump, and the end of the response i get:
> >>
> >
> > What does 'ceph osd tree' tell you?
> >
> > It seems there is something wrong with your CRUSHMap.
> >
> > Wido
> >
> >
> > Thanks for your answer Wido.
> >
> > Here is the output of ceph osd tree;
> >
> > # ceph osd tree
> > ID WEIGHT TYPE NAME       UP/DOWN REWEIGHT PRIMARY-AFFINITY
> > -1      0 root default
> > -2      0     host ceph01
> >  0      0         osd.0        up  1.00000          1.00000
> >  3      0         osd.3        up  1.00000          1.00000
> >  6      0         osd.6        up  1.00000          1.00000
> > -3      0     host ceph02
> >  1      0         osd.1        up  1.00000          1.00000
> >  4      0         osd.4        up  1.00000          1.00000
> >  7      0         osd.7        up  1.00000          1.00000
> > -4      0     host ceph03
> >  2      0         osd.2        up  1.00000          1.00000
> >  5      0         osd.5        up  1.00000          1.00000
> >  8      0         osd.8        up  1.00000          1.00000
> >
> >
>
> The weights are allo zero (0) of all the OSDs. How big are the disks? I
> think they are very tiny , eg <10GB?
>
> You probably want a bit bigger disks to test with.
>
> Or set the weight manually of each OSD:
>
> $ ceph osd crush reweight osd.X 1
>
> Wido
>
> >>
> >> osdstatkbusedkbavailkbhb inhb out
> >> 03668851949085231596[1,2,3,4,5,6,7,8][]
> >> 13400451975925231596[][]
> >> 23400451975925231596[1][]
> >> 33400451975925231596[0,1,2,4,5,6,7,8][]
> >> 43400451975925231596[1,2][]
> >> 53400451975925231596[1,2,4][]
> >> 63400451975925231596[0,1,2,3,4,5,7,8][]
> >> 73400451975925231596[1,2,4,5][]
> >> 83400451975925231596[1,2,4,5,7][]
> >>  sum3087204677564447084364
> >>
> >>
> >> Please someone can help me?
> >>
> >>
> >>
> >> 2015-07-13 11:45 GMT+02:00 alberto ayllon <albertoayllonces at
> > gmail.com <http://gmail.com>
> >> <mailto:albertoayllonces <mailto:albertoayllonces> at gmail.com
> > <http://gmail.com>>>:
> >>
> >>     Hello everybody and thanks foryour help.
> >>
> >>     Hello, I'm newbie in CEPH, I'm trying to install a CEPHcluster with
> >>     test purpose.
> >>
> >>     I had just installed a CEPH cluster with three VMs (ubuntu 14.04),
> >>     each one has one mon daemon and three OSDs, also each server has 3
> > disk.
> >>     Cluster has only one poll (rbd) with pg and pgp_num = 280, and "osd
> >>     pool get rbd size = 2".
> >>
> >>     I made cluster's installation with  ceph-deploy, ceph version is
> >>     "0.94.2"
> >>
> >>     I think cluster's OSDs are having peering problems, because if Irun
> >>     ceph status, it returns:
> >>
> >>     # ceph status
> >>         cluster d54a2216-b522-4744-a7cc-a2106e1281b6
> >>          health HEALTH_WARN
> >>                 280 pgs degraded
> >>                 280 pgs stuck degraded
> >>                 280 pgs stuck unclean
> >>                 280 pgs stuck undersized
> >>                 280 pgs undersized
> >>          monmap e3: 3 mons at
> >>
> > {ceph01=
> 172.16.70.158:6789/0,ceph02=172.16.70.159:6789/0,ceph03=172.16.70.160:6789/0
> > <
> http://172.16.70.158:6789/0,ceph02=172.16.70.159:6789/0,ceph03=172.16.70.160:6789/0
> >
> >>
> > <
> http://172.16.70.158:6789/0,ceph02=172.16.70.159:6789/0,ceph03=172.16.70.160:6789/0
> >}
> >>                 election epoch 38, quorum 0,1,2 ceph01,ceph02,ceph03
> >>          osdmap e46: 9 osds: 9 up, 9 in
> >>           pgmap v129: 280 pgs, 1 pools, 0 bytes data, 0 objects
> >>                 301 MB used, 45679 MB / 45980 MB avail
> >>                      280 active+undersized+degraded
> >>
> >>     And for all pgs, the command "ceph pg map X.yy"returns something
> like:
> >>
> >>     osdmap e46 pg 0.d7 (0.d7) -> up [0] acting [0]
> >>
> >>     As I know "Acting Set" and "Up Set" must have the same value, but as
> >>     they are equal to 0, there are not defined OSDs to
> >>     stores pgs replicas, and I think this is why all pg are in
> >>     "active+undersized+degraded" state.
> >>
> >>     Has anyone any idea of what I have to do for  "Active Set" and "Up
> >>     Set" reaches correct values.
> >>
> >>
> >>     Thanks a lot!
> >>
> >>
> >>
> >>
> >> _______________________________________________
> >> ceph-users mailing list
> >> ceph-users at lists.ceph.com <http://lists.ceph.com>
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to