Your CRUSH rule specifies to select 3 different chassis but your CRUSH
ma defines no chassis.
Add buckets of type chassis or change the rule to select hosts.

Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90

On Tue, Jan 29, 2019 at 7:40 PM PHARABOT Vincent
<vincent.phara...@3ds.com> wrote:
>
> Sorry JC, here is the correct osd crush rule dump (type=chassis instead of 
> host)
>
>
>
> # ceph osd crush rule dump
>
> [
>
> {
>
> "rule_id": 0,
>
> "rule_name": "replicated_rule",
>
> "ruleset": 0,
>
> "type": 1,
>
> "min_size": 1,
>
> "max_size": 10,
>
> "steps": [
>
> {
>
> "op": "take",
>
> "item": -1,
>
> "item_name": "default"
>
> },
>
> {
>
> "op": "chooseleaf_firstn",
>
> "num": 0,
>
> "type": "chassis"
>
> },
>
> {
>
> "op": "emit"
>
> }
>
> ]
>
> }
>
> ]
>
>
>
> De : ceph-users [mailto:ceph-users-boun...@lists.ceph.com] De la part de 
> PHARABOT Vincent
> Envoyé : mardi 29 janvier 2019 19:33
> À : Jean-Charles Lopez <jelo...@redhat.com>
> Cc : ceph-users@lists.ceph.com
> Objet : Re: [ceph-users] Bright new cluster get all pgs stuck in inactive
>
>
>
> Thanks for the quick reply
>
>
>
> Here is the result
>
>
>
> # ceph osd crush rule dump
>
> [
>
>     {
>
>         "rule_id": 0,
>
>         "rule_name": "replicated_rule",
>
>         "ruleset": 0,
>
>         "type": 1,
>
>         "min_size": 1,
>
>         "max_size": 10,
>
>         "steps": [
>
>             {
>
>                 "op": "take",
>
>                 "item": -1,
>
>                 "item_name": "default"
>
>             },
>
>             {
>
>                 "op": "chooseleaf_firstn",
>
>                 "num": 0,
>
>                 "type": "host"
>
>             },
>
>             {
>
>                "op": "emit"
>
>             }
>
>         ]
>
>     }
>
> ]
>
>
>
> De : Jean-Charles Lopez [mailto:jelo...@redhat.com]
> Envoyé : mardi 29 janvier 2019 19:30
> À : PHARABOT Vincent <vincent.phara...@3ds.com>
> Cc : ceph-users@lists.ceph.com
> Objet : Re: [ceph-users] Bright new cluster get all pgs stuck in inactive
>
>
>
> Hi,
>
>
>
> I suspect your generated CRUSH rule is incorret because of 
> osd_crush_cooseleaf_type=2 and by default chassis bucket are not created.
>
>
>
> Changing the type of bucket to host (osd_crush_cooseleaf_type=1 which is the 
> default when using old ceph-deploy or ceph-ansible) for your deployment 
> should fix the problem.
>
>
>
> Could you show the output of ceph osd crush rule dump to verify how the rule 
> was built
>
>
>
> JC
>
>
>
> On Jan 29, 2019, at 10:08, PHARABOT Vincent <vincent.phara...@3ds.com> wrote:
>
>
>
> Hello,
>
>
>
> I have a bright new cluster with 2 pools, but cluster keeps pgs in inactive 
> state.
>
> I have 3 OSDs and 1 Mon… all seems ok except I could not have pgs in 
> clean+active state !
>
>
>
> I might miss something obvious but I really don’t know what…. Someone could 
> help me ?
>
> I tried to seek answers among the list mail threads, but no luck, other 
> situation seems different
>
>
>
> Thank you a lot for your help
>
>
>
> Vincent
>
>
>
> # ceph -v
>
> ceph version 13.2.4 (b10be4d44915a4d78a8e06aa31919e74927b142e) mimic (stable)
>
>
>
> # ceph -s
>
> cluster:
>
> id: ff4c91fb-3c29-4d9f-a26f-467d6b6a712e
>
> health: HEALTH_WARN
>
> Reduced data availability: 200 pgs inactive
>
>
>
> services:
>
> mon: 1 daemons, quorum ip-10-8-66-123.eu-west-2.compute.internal
>
> mgr: ip-10-8-66-123.eu-west-2.compute.internal(active)
>
> osd: 3 osds: 3 up, 3 in
>
>
>
> data:
>
> pools: 2 pools, 200 pgs
>
> objects: 0 objects, 0 B
>
> usage: 3.0 GiB used, 2.9 TiB / 2.9 TiB avail
>
> pgs: 100.000% pgs unknown
>
> 200 unknown
>
>
>
> # ceph osd tree -f json-pretty
>
>
>
> {
>
>     "nodes": [
>
>         {
>
>             "id": -1,
>
>             "name": "default",
>
>             "type": "root",
>
>             "type_id": 10,
>
>             "children": [
>
>                 -3,
>
>                 -5,
>
>                 -7
>
>             ]
>
>         },
>
>         {
>
>             "id": -7,
>
>             "name": "ip-10-8-10-108",
>
>             "type": "host",
>
>             "type_id": 1,
>
>             "pool_weights": {},
>
>             "children": [
>
>                 2
>
>             ]
>
>         },
>
>         {
>
>             "id": 2,
>
>             "device_class": "hdd",
>
>             "name": "osd.2",
>
>             "type": "osd",
>
>             "type_id": 0,
>
>             "crush_weight": 0.976593,
>
>             "depth": 2,
>
>             "pool_weights": {},
>
>             "exists": 1,
>
>             "status": "up",
>
>             "reweight": 1.000000,
>
>             "primary_affinity": 1.000000
>
>         },
>
>         {
>
>             "id": -5,
>
>             "name": "ip-10-8-22-148",
>
>             "type": "host",
>
>             "type_id": 1,
>
>             "pool_weights": {},
>
>             "children": [
>
>                 1
>
>             ]
>
>         },
>
>         {
>
>             "id": 1,
>
>             "device_class": "hdd",
>
>             "name": "osd.1",
>
>             "type": "osd",
>
>             "type_id": 0,
>
>             "crush_weight": 0.976593,
>
>             "depth": 2,
>
>             "pool_weights": {},
>
>             "exists": 1,
>
>             "status": "up",
>
>             "reweight": 1.000000,
>
>             "primary_affinity": 1.000000
>
>         },
>
>         {
>
>             "id": -3,
>
>             "name": "ip-10-8-5-246",
>
>             "type": "host",
>
>             "type_id": 1,
>
>             "pool_weights": {},
>
>             "children": [
>
>                 0
>
>             ]
>
>         },
>
>         {
>
>             "id": 0,
>
>             "device_class": "hdd",
>
>             "name": "osd.0",
>
>             "type": "osd",
>
>             "type_id": 0,
>
>             "crush_weight": 0.976593,
>
>            "depth": 2,
>
>             "pool_weights": {},
>
>             "exists": 1,
>
>             "status": "up",
>
>             "reweight": 1.000000,
>
>             "primary_affinity": 1.000000
>
>         }
>
>     ],
>
>     "stray": []
>
> }
>
>
>
> # cat /etc/ceph/ceph.conf
>
> [global]
>
> fsid = ff4c91fb-3c29-4d9f-a26f-467d6b6a712e
>
> mon initial members = ip-10-8-66-123
>
> mon host = 10.8.66.123
>
> auth_cluster_required = cephx
>
> auth_service_required = cephx
>
> auth_client_required = cephx
>
> pid file = /var/run/$cluster/$type.pid
>
>
>
>
>
> #Choose reasonable numbers for number of replicas and placement groups.
>
> osd pool default size = 3 # Write an object 3 times
>
> osd pool default min size = 2 # Allow writing 2 copy in a degraded state
>
> osd pool default pg num = 100
>
> osd pool default pgp num = 100
>
>
>
> #Choose a reasonable crush leaf type
>
> #0 for a 1-node cluster.
>
> #1 for a multi node cluster in a single rack
>
> #2 for a multi node, multi chassis cluster with multiple hosts in a chassis
>
> #3 for a multi node cluster with hosts across racks, etc.
>
> osd crush chooseleaf type = 2
>
>
>
> [mon]
>
>         debug mon = 20
>
>
>
> # ceph health detail
>
> HEALTH_WARN Reduced data availability: 200 pgs inactive
>
> PG_AVAILABILITY Reduced data availability: 200 pgs inactive
>
>     pg 1.46 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.47 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.48 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.49 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.4a is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.4b is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.4c is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.4d is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.4e is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.4f is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.50 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.51 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.52 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.53 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.54 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.55 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.56 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.57 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.58 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.59 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.5a is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.5b is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.5c is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.5d is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.5e is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 1.5f is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>
>     pg 2.44 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.48 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.49 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.4a is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.4b is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.4c is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.4d is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.4e is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.4f is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.50 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.51 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.52 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.53 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.54 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.55 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.56 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.57 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.58 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.59 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.5a is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.5b is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.5c is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.5d is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.5e is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
>     pg 2.5f is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>
> This email and any attachments are intended solely for the use of the 
> individual or entity to whom it is addressed and may be confidential and/or 
> privileged.
>
> If you are not one of the named recipients or have received this email in 
> error,
>
> (i) you should not read, disclose, or copy it,
>
> (ii) please notify sender of your receipt by reply email and delete this 
> email and all attachments,
>
> (iii) Dassault Systèmes does not accept or assume any liability or 
> responsibility for any use of or reliance on this email.
>
>
>
> Please be informed that your personal data are processed according to our 
> data privacy policy as described on our website. Should you have any 
> questions related to personal data protection, please contact 3DS Data 
> Protection Officer at 3ds.compliance-priv...@3ds.com
>
>
>
> For other languages, go to https://www.3ds.com/terms/email-disclaimer
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> This email and any attachments are intended solely for the use of the 
> individual or entity to whom it is addressed and may be confidential and/or 
> privileged.
>
> If you are not one of the named recipients or have received this email in 
> error,
>
> (i) you should not read, disclose, or copy it,
>
> (ii) please notify sender of your receipt by reply email and delete this 
> email and all attachments,
>
> (iii) Dassault Systèmes does not accept or assume any liability or 
> responsibility for any use of or reliance on this email.
>
>
>
> Please be informed that your personal data are processed according to our 
> data privacy policy as described on our website. Should you have any 
> questions related to personal data protection, please contact 3DS Data 
> Protection Officer at 3ds.compliance-priv...@3ds.com
>
>
>
> For other languages, go to https://www.3ds.com/terms/email-disclaimer
>
> This email and any attachments are intended solely for the use of the 
> individual or entity to whom it is addressed and may be confidential and/or 
> privileged.
>
> If you are not one of the named recipients or have received this email in 
> error,
>
> (i) you should not read, disclose, or copy it,
>
> (ii) please notify sender of your receipt by reply email and delete this 
> email and all attachments,
>
> (iii) Dassault Systèmes does not accept or assume any liability or 
> responsibility for any use of or reliance on this email.
>
>
> Please be informed that your personal data are processed according to our 
> data privacy policy as described on our website. Should you have any 
> questions related to personal data protection, please contact 3DS Data 
> Protection Officer at 3ds.compliance-priv...@3ds.com
>
>
> For other languages, go to https://www.3ds.com/terms/email-disclaimer
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to