Hi,

I suspect your generated CRUSH rule is incorret because of 
osd_crush_cooseleaf_type=2 and by default chassis bucket are not created.

Changing the type of bucket to host (osd_crush_cooseleaf_type=1 which is the 
default when using old ceph-deploy or ceph-ansible) for your deployment should 
fix the problem.

Could you show the output of ceph osd crush rule dump to verify how the rule 
was built

JC

> On Jan 29, 2019, at 10:08, PHARABOT Vincent <vincent.phara...@3ds.com> wrote:
> 
> Hello,
>  
> I have a bright new cluster with 2 pools, but cluster keeps pgs in inactive 
> state.
> I have 3 OSDs and 1 Mon… all seems ok except I could not have pgs in 
> clean+active state !
>  
> I might miss something obvious but I really don’t know what…. Someone could 
> help me ?
> I tried to seek answers among the list mail threads, but no luck, other 
> situation seems different
>  
> Thank you a lot for your help
>  
> Vincent
>  
> # ceph -v
> ceph version 13.2.4 (b10be4d44915a4d78a8e06aa31919e74927b142e) mimic (stable)
>  
> # ceph -s
> cluster:
> id: ff4c91fb-3c29-4d9f-a26f-467d6b6a712e
> health: HEALTH_WARN
> Reduced data availability: 200 pgs inactive
>  
> services:
> mon: 1 daemons, quorum ip-10-8-66-123.eu 
> <http://ip-10-8-66-123.eu/>-west-2.compute.internal
> mgr: ip-10-8-66-123.eu 
> <http://ip-10-8-66-123.eu/>-west-2.compute.internal(active)
> osd: 3 osds: 3 up, 3 in
>  
> data:
> pools: 2 pools, 200 pgs
> objects: 0 objects, 0 B
> usage: 3.0 GiB used, 2.9 TiB / 2.9 TiB avail
> pgs: 100.000% pgs unknown
> 200 unknown
>  
> # ceph osd tree -f json-pretty
>  
> {
>     "nodes": [
>         {
>             "id": -1,
>             "name": "default",
>             "type": "root",
>             "type_id": 10,
>             "children": [
>                 -3,
>                 -5,
>                 -7
>             ]
>         },
>         {
>             "id": -7,
>             "name": "ip-10-8-10-108",
>             "type": "host",
>             "type_id": 1,
>             "pool_weights": {},
>             "children": [
>                 2
>             ]
>         },
>         {
>             "id": 2,
>             "device_class": "hdd",
>             "name": "osd.2",
>             "type": "osd",
>             "type_id": 0,
>             "crush_weight": 0.976593,
>             "depth": 2,
>             "pool_weights": {},
>             "exists": 1,
>             "status": "up",
>             "reweight": 1.000000,
>             "primary_affinity": 1.000000
>         },
>         {
>             "id": -5,
>             "name": "ip-10-8-22-148",
>             "type": "host",
>             "type_id": 1,
>             "pool_weights": {},
>             "children": [
>                 1
>             ]
>         },
>         {
>             "id": 1,
>             "device_class": "hdd",
>             "name": "osd.1",
>             "type": "osd",
>             "type_id": 0,
>             "crush_weight": 0.976593,
>             "depth": 2,
>             "pool_weights": {},
>             "exists": 1,
>             "status": "up",
>             "reweight": 1.000000,
>             "primary_affinity": 1.000000
>         },
>         {
>             "id": -3,
>             "name": "ip-10-8-5-246",
>             "type": "host",
>             "type_id": 1,
>             "pool_weights": {},
>             "children": [
>                 0
>             ]
>         },
>         {
>             "id": 0,
>             "device_class": "hdd",
>             "name": "osd.0",
>             "type": "osd",
>             "type_id": 0,
>             "crush_weight": 0.976593,
>            "depth": 2,
>             "pool_weights": {},
>             "exists": 1,
>             "status": "up",
>             "reweight": 1.000000,
>             "primary_affinity": 1.000000
>         }
>     ],
>     "stray": []
> }
>  
> # cat /etc/ceph/ceph.conf
> [global]
> fsid = ff4c91fb-3c29-4d9f-a26f-467d6b6a712e
> mon initial members = ip-10-8-66-123
> mon host = 10.8.66.123
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
> pid file = /var/run/$cluster/$type.pid
>  
>  
> #Choose reasonable numbers for number of replicas and placement groups.
> osd pool default size = 3 # Write an object 3 times
> osd pool default min size = 2 # Allow writing 2 copy in a degraded state
> osd pool default pg num = 100
> osd pool default pgp num = 100
>  
> #Choose a reasonable crush leaf type
> #0 for a 1-node cluster.
> #1 for a multi node cluster in a single rack
> #2 for a multi node, multi chassis cluster with multiple hosts in a chassis
> #3 for a multi node cluster with hosts across racks, etc.
> osd crush chooseleaf type = 2
>  
> [mon]
>         debug mon = 20
>  
> # ceph health detail
> HEALTH_WARN Reduced data availability: 200 pgs inactive
> PG_AVAILABILITY Reduced data availability: 200 pgs inactive
>     pg 1.46 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.47 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.48 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.49 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.4a is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.4b is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.4c is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.4d is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.4e is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.4f is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.50 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.51 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.52 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.53 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.54 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.55 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.56 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.57 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.58 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.59 is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.5a is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.5b is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.5c is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.5d is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.5e is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 1.5f is stuck inactive for 10848.068201, current state unknown, last 
> acting []
>     pg 2.44 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.48 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.49 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.4a is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.4b is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.4c is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.4d is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.4e is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.4f is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.50 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.51 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.52 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.53 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.54 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.55 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.56 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.57 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.58 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.59 is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.5a is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.5b is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.5c is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.5d is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.5e is stuck inactive for 10846.400420, current state unknown, last 
> acting []
>     pg 2.5f is stuck inactive for 10846.400420, current state unknown, last 
> acting []
> This email and any attachments are intended solely for the use of the 
> individual or entity to whom it is addressed and may be confidential and/or 
> privileged.
> If you are not one of the named recipients or have received this email in 
> error,
> (i) you should not read, disclose, or copy it,
> (ii) please notify sender of your receipt by reply email and delete this 
> email and all attachments,
> (iii) Dassault Systèmes does not accept or assume any liability or 
> responsibility for any use of or reliance on this email.
> 
> Please be informed that your personal data are processed according to our 
> data privacy policy as described on our website. Should you have any 
> questions related to personal data protection, please contact 3DS Data 
> Protection Officer at 3ds.compliance-priv...@3ds.com 
> <mailto:3ds.compliance-priv...@3ds.com>
> 
> For other languages, go to https://www.3ds.com/terms/email-disclaimer 
> <https://www.3ds.com/terms/email-disclaimer>_______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to