Nope, all osds are running 0.94.9

On 28/03/17 14:53, Brian Andrus wrote:
Well, you said you were running v0.94.9, but are there any OSDs running pre-v0.94.4 as the error states?

On Tue, Mar 28, 2017 at 6:51 AM, Jaime Ibar <ja...@tchpc.tcd.ie <mailto:ja...@tchpc.tcd.ie>> wrote:



    On 28/03/17 14:41, Brian Andrus wrote:
    What does
    # ceph tell osd.* version
    ceph tell osd.21 version
    Error ENXIO: problem getting command descriptions from osd.21

    reveal? Any pre-v0.94.4 hammer OSDs running as the error states?
    Yes, as this is the first one I tried to upgrade.
    The other ones are running hammer

    Thanks



    On Tue, Mar 28, 2017 at 1:21 AM, Jaime Ibar <ja...@tchpc.tcd.ie
    <mailto:ja...@tchpc.tcd.ie>> wrote:

        Hi,

        I did change the ownership to user ceph. In fact, OSD
        processes are running

        ps aux | grep ceph
ceph 2199 0.0 2.7 1729044 918792 ? Ssl Mar27 0:21 /usr/bin/ceph-osd --cluster=ceph -i 42 -f --setuser
        ceph --setgroup ceph
ceph 2200 0.0 2.7 1721212 911084 ? Ssl Mar27 0:20 /usr/bin/ceph-osd --cluster=ceph -i 18 -f --setuser
        ceph --setgroup ceph
ceph 2212 0.0 2.8 1732532 926580 ? Ssl Mar27 0:20 /usr/bin/ceph-osd --cluster=ceph -i 3 -f --setuser ceph
        --setgroup ceph
ceph 2215 0.0 2.8 1743552 935296 ? Ssl Mar27 0:20 /usr/bin/ceph-osd --cluster=ceph -i 47 -f --setuser
        ceph --setgroup ceph
ceph 2341 0.0 2.7 1715548 908312 ? Ssl Mar27 0:20 /usr/bin/ceph-osd --cluster=ceph -i 51 -f --setuser
        ceph --setgroup ceph
ceph 2383 0.0 2.7 1694944 893768 ? Ssl Mar27 0:20 /usr/bin/ceph-osd --cluster=ceph -i 56 -f --setuser
        ceph --setgroup ceph
        [...]

        If I run one of the osd increasing debug

        ceph-osd --debug_osd 5 -i 31

        this is what I get in logs

        [...]

        0 osd.31 14016 done with init, starting boot process
        2017-03-28 09:19:15.280182 7f083df0c800  1 osd.31 14016 We
        are healthy, booting
        2017-03-28 09:19:15.280685 7f081cad3700  1 osd.31 14016
        osdmap indicates one or more pre-v0.94.4 hammer OSDs is running
        [...]

        It seems the osd is running but ceph is not aware of it

        Thanks
        Jaime




        On 27/03/17 21:56, George Mihaiescu wrote:

            Make sure the OSD processes on the Jewel node are
            running. If you didn't change the ownership to user ceph,
            they won't start.


                On Mar 27, 2017, at 11:53, Jaime Ibar
                <ja...@tchpc.tcd.ie <mailto:ja...@tchpc.tcd.ie>> wrote:

                Hi all,

                I'm upgrading ceph cluster from Hammer 0.94.9 to
                jewel 10.2.6.

                The ceph cluster has 3 servers (one mon and one mds
                each) and another 6 servers with
                12 osds each.
                The monitoring and mds have been succesfully upgraded
                to latest jewel release, however
                after upgrade the first osd server(12 osds), ceph is
                not aware of them and
                are marked as down

                ceph -s

                cluster 4a158d27-f750-41d5-9e7f-26ce4c9d2d45
                     health HEALTH_WARN
                [...]
                            12/72 in osds are down
                            noout flag(s) set
                     osdmap e14010: 72 osds: 60 up, 72 in; 14641
                remapped pgs
                            flags noout
                [...]

                ceph osd tree

                3   3.64000         osd.3 down  1.00000 1.00000
                8   3.64000         osd.8 down  1.00000 1.00000
                14   3.64000         osd.14  down  1.00000 1.00000
18 3.64000 osd.18 down 1.00000 1.00000 21 3.64000 osd.21 down 1.00000 1.00000 28 3.64000 osd.28 down 1.00000 1.00000 31 3.64000 osd.31 down 1.00000 1.00000 37 3.64000 osd.37 down 1.00000 1.00000 42 3.64000 osd.42 down 1.00000 1.00000 47 3.64000 osd.47 down 1.00000 1.00000 51 3.64000 osd.51 down 1.00000 1.00000 56 3.64000 osd.56 down 1.00000 1.00000

                If I run this command with one of the down osd
                ceph osd in 14
                osd.14 is already in.
                however ceph doesn't mark it as up and the cluster
                health remains
                in degraded state.

                Do I have to upgrade all the osds to jewel first?
                Any help as I'm running out of ideas?

                Thanks
                Jaime

--
                Jaime Ibar
                High Performance & Research Computing, IS Services
                Lloyd Building, Trinity College Dublin, Dublin 2,
                Ireland.
                http://www.tchpc.tcd.ie/ | ja...@tchpc.tcd.ie
                <mailto:ja...@tchpc.tcd.ie>
                Tel: +353-1-896-3725 <tel:%2B353-1-896-3725>

                _______________________________________________
                ceph-users mailing list
                ceph-users@lists.ceph.com
                <mailto:ceph-users@lists.ceph.com>
                http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
                <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>


--
        Jaime Ibar
        High Performance & Research Computing, IS Services
        Lloyd Building, Trinity College Dublin, Dublin 2, Ireland.
        http://www.tchpc.tcd.ie/ | ja...@tchpc.tcd.ie
        <mailto:ja...@tchpc.tcd.ie>
        Tel: +353-1-896-3725 <tel:%2B353-1-896-3725>

        _______________________________________________
        ceph-users mailing list
        ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
        http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
        <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>




-- Brian Andrus | Cloud Systems Engineer | DreamHost
    brian.and...@dreamhost.com <mailto:brian.and...@dreamhost.com> |
    www.dreamhost.com <http://www.dreamhost.com>

--
    Jaime Ibar
    High Performance & Research Computing, IS Services
    Lloyd Building, Trinity College Dublin, Dublin 2, Ireland.
    http://www.tchpc.tcd.ie/  |ja...@tchpc.tcd.ie <mailto:ja...@tchpc.tcd.ie>
    Tel:+353-1-896-3725 <tel:+353%201%20896%203725>




--
Brian Andrus | Cloud Systems Engineer | DreamHost
brian.and...@dreamhost.com | www.dreamhost.com <http://www.dreamhost.com>

--

Jaime Ibar
High Performance & Research Computing, IS Services
Lloyd Building, Trinity College Dublin, Dublin 2, Ireland.
http://www.tchpc.tcd.ie/ | ja...@tchpc.tcd.ie
Tel: +353-1-896-3725

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to