[ceph-users] Quincy: Stuck on image permissions

2023-02-12 Thread hicks
Hello guys,

could someone help me with this? We've been long-time CEPH users... runing 
several Mimic + Pacific CEPH clusters. Dozens of disk per cluster, typically.

BUT... now I have this brand new Quincy cluster and I'm not able to give CLIENT 
(Quincy on Rocky 8) rw access to ONE IMAGE on Quincy cluster (cephadm / Rocky 
9).

I'm using something what worked for us for ages:

rbd auth ls:
client.xxx
key: ...
caps: [mon] profile rbd
caps: [osd] allow rwx pool prod object_prefix rbd_data.600d1c6723ae; 
allow rwx pool prod object_prefix rbd_header.600d1c6723ae; allow rx pool prod 
object_prefix rbd_id.xxx-data

rbd info:
rbd image 'xxx-data':
size 2 TiB in 524288 objects
order 22 (4 MiB objects)
snapshot_count: 2
id: 600d1c6723ae
block_name_prefix: rbd_data.600d1c6723ae
format: 2
features: layering, exclusive-lock, object-map, fast-diff, deep-flatten
op_features:
flags:

rados ls:
rbd_data.600d1c6723ae.0003958d
rbd_header.600d1c6723ae
rbd_id.xxx-data

BUT... it DOES NOT WORK. When I try it to map on client it says:

2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::image::GetMetadataRequest: 
0x7f3a1c001f40 handle_metadata_list: failed to retrieve image metadata: (1) 
Operation not permitted
2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::image::RefreshRequest: 
failed to retrieve pool metadata: (1) Operation not permitted
2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::image::OpenRequest: failed 
to refresh image: (1) Operation not permitted
2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::ImageState: 0x555eff78cfc0 
failed to open image: (1) Operation not permitted
rbd: error opening image xxx-data: (1) Operation not permitted

The mapping and access DOES work when I put "osd allow *" into ceph auth.

What is the recommended syntax for Quincy? 

btw: this use case should be mentioned in the manual I think...

Thanks!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph Quincy On Rocky 8.x - Upgrade To Rocky 9.1

2023-02-12 Thread hicks
We're running Quincy cluster on Rocky9... its in the podman and also, you can 
install ceph-common (Quincy version) from packages.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Extremally need help. Openshift cluster is down :c

2023-02-12 Thread kreept . sama
Hello Eugen, yes i have
Its from object a
...
debug 2023-02-12T07:12:55.469+ 7f66af51e700  1 mds.gml-okd-cephfs-a 
asok_command: status {prefix=status} (starting...)
debug 2023-02-12T07:13:05.453+ 7f66af51e700  1 mds.gml-okd-cephfs-a 
asok_command: status {prefix=status} (starting...)
debug 2023-02-12T07:13:15.478+ 7f66af51e700  1 mds.gml-okd-cephfs-a 
asok_command: status {prefix=status} (starting...)
debug 2023-02-12T07:13:25.477+ 7f66af51e700  1 mds.gml-okd-cephfs-a 
asok_command: status {prefix=status} (starting...)
debug 2023-02-12T07:13:35.445+ 7f66af51e700  1 mds.gml-okd-cephfs-a 
asok_command: status {prefix=status} (starting...)
debug 2023-02-12T07:13:45.487+ 7f66af51e700  1 mds.gml-okd-cephfs-a 
asok_command: status {prefix=status} (starting...)
...
and the same for object b

...
debug 2023-02-12T07:15:41.496+ 7fdf6281e700  1 mds.gml-okd-cephfs-b 
asok_command: status {prefix=status} (starting...)
debug 2023-02-12T07:15:51.479+ 7fdf6281e700  1 mds.gml-okd-cephfs-b 
asok_command: status {prefix=status} (starting...)
debug 2023-02-12T07:16:01.477+ 7fdf6281e700  1 mds.gml-okd-cephfs-b 
asok_command: status {prefix=status} (starting...)
...
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Quincy: Stuck on image permissions

2023-02-12 Thread Jakub Chromy

Hello,

looks like I've found it -- THE NAMESPACES :)

I love it. Thanks!


On 11/02/2023 21:37, hi...@cgi.cz wrote:

Hello guys,

could someone help me with this? We've been long-time CEPH users... runing 
several Mimic + Pacific CEPH clusters. Dozens of disk per cluster, typically.

BUT... now I have this brand new Quincy cluster and I'm not able to give CLIENT 
(Quincy on Rocky 8) rw access to ONE IMAGE on Quincy cluster (cephadm / Rocky 
9).

I'm using something what worked for us for ages:

rbd auth ls:
client.xxx
 key: ...
 caps: [mon] profile rbd
 caps: [osd] allow rwx pool prod object_prefix rbd_data.600d1c6723ae; 
allow rwx pool prod object_prefix rbd_header.600d1c6723ae; allow rx pool prod 
object_prefix rbd_id.xxx-data

rbd info:
rbd image 'xxx-data':
 size 2 TiB in 524288 objects
 order 22 (4 MiB objects)
 snapshot_count: 2
 id: 600d1c6723ae
 block_name_prefix: rbd_data.600d1c6723ae
 format: 2
 features: layering, exclusive-lock, object-map, fast-diff, deep-flatten
 op_features:
 flags:

rados ls:
rbd_data.600d1c6723ae.0003958d
rbd_header.600d1c6723ae
rbd_id.xxx-data

BUT... it DOES NOT WORK. When I try it to map on client it says:

2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::image::GetMetadataRequest: 
0x7f3a1c001f40 handle_metadata_list: failed to retrieve image metadata: (1) 
Operation not permitted
2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::image::RefreshRequest: 
failed to retrieve pool metadata: (1) Operation not permitted
2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::image::OpenRequest: failed 
to refresh image: (1) Operation not permitted
2023-02-11T20:49:18.665+0100 7f3a337fe700 -1 librbd::ImageState: 0x555eff78cfc0 
failed to open image: (1) Operation not permitted
rbd: error opening image xxx-data: (1) Operation not permitted

The mapping and access DOES work when I put "osd allow *" into ceph auth.

What is the recommended syntax for Quincy?

btw: this use case should be mentioned in the manual I think...

Thanks!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Subject: OSDs added, remapped pgs and objects misplaced cycling up and down

2023-02-12 Thread Chris Dunlop

Hi,

ceph-16.2.9

I've added some new osds - some added to existing hosts and some on 
newly-commissioned hosts. The new osds were added to the data side of an 
existing EC 8.3 pool.


I've been waiting for the system to finish remapping / backfilling for 
some time. Originally the number of remapped pgs and objects misplaced was 
steadily decreasing. However in the last few days both remapped pgs and 
objects misplaced have increased several times. Both then steadily 
decrease before jumping up again. See below

for some details on this.

I'm guessing these are associated stats, i.e. the number of objects 
misplaced reflects the objects in the remapped pgs so it's not surprising 
they're moving in concert.


I thought at first that this cycling might be due to the balancer (which 
was enabled), but the cycling is continuing after the balancer was 
disabled, e.g.:


$ ts; ceph balancer status
2023-02-09 12:39:14
{
"active": true,
"last_optimize_duration": "0:00:00.000313",
"last_optimize_started": "Thu Feb  9 01:38:55 2023",
"mode": "upmap",
"optimize_result": "Too many objects (0.078537 > 0.05) are misplaced; try 
again later",
"plans": []
}
$ ts; ceph balancer off
2023-02-09 13:03:27 ceph balancer off

4 days later, the balancer is still inactive, and "last_optimize_started" 
indicates it hasn't run since before it was turned off:


$ ts; ceph balancer status
2023-02-13 10:18:18
{
"active": false,
"last_optimize_duration": "0:00:00.000420",
"last_optimize_started": "Thu Feb  9 02:02:56 2023",
"mode": "upmap",
"optimize_result": "Too many objects (0.078101 > 0.05) are misplaced; try 
again later",
"plans": []
}

Since the balancer was turned off the cycling continues: "misplaced" gets 
down to 5% and then rapidly increases to 9%, in conjunction with "starting 
backfill" messages, before starting to slowly decrease once again. See 
below for a log extract.


Is this "sawtooth" pattern of remapped pgs and misplaced objects a normal 
consequence of adding OSDs?


Will it eventually settle done itself or do I need to "do something"?

Is there some way of telling how much work remains until it all settles 
down?


Cheers,

Chris

--
#
# General health
#   scrubs disabled until added OSDs complete backfilling etc.
#
$ ceph -s
  cluster:
id: c6618970-0ce0-4cb2-bc9a-dd5f29b62e24
health: HEALTH_WARN
noout,noscrub,nodeep-scrub flag(s) set
5945 pgs not deep-scrubbed in time
5945 pgs not scrubbed in time
5 slow ops, oldest one blocked for 464 sec, mon.k2 has slow ops
 
  services:

mon: 5 daemons, quorum k2,b2,b4,k1,b5 (age 5d)
mgr: b4(active, since 3M), standbys: b5, b2
osd: 82 osds: 82 up (since 2w), 82 in (since 2w); 12 remapped pgs
 flags noout,noscrub,nodeep-scrub
 
  data:

pools:   16 pools, 5945 pgs
objects: 70.35M objects, 88 TiB
usage:   186 TiB used, 242 TiB / 428 TiB avail
pgs: 50050192/700705500 objects misplaced (7.143%)
 5933 active+clean
 10   active+remapped+backfill_wait
 2active+remapped+backfilling
 
  io:

client:   2.2 MiB/s rd, 1.8 MiB/s wr, 395 op/s rd, 122 op/s wr
recovery: 24 MiB/s, 24 objects/s

#
# remapped pgs steadily decreasing then started periodically increasing
#   (extract from periodic "ceph -s")
#
2023-02-06 00:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 26 
remapped pgs
2023-02-06 04:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 25 
remapped pgs
2023-02-06 06:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 24 
remapped pgs
2023-02-06 09:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d); 23 
remapped pgs
2023-02-06 12:30:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 22 
remapped pgs
2023-02-06 14:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 21 
remapped pgs
2023-02-06 14:30:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 20 
remapped pgs
2023-02-06 20:30:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d); 19 
remapped pgs
2023-02-07 00:00:00osd: 82 osds: 82 up (since 8d), 82 in (since 7d); 18 
remapped pgs
2023-02-07 01:00:00osd: 82 osds: 82 up (since 8d), 82 in (since 7d); 17 
remapped pgs
2023-02-07 11:00:00osd: 82 osds: 82 up (since 8d), 82 in (since 8d); 15 
remapped pgs
2023-02-07 16:30:00osd: 82 osds: 82 up (since 8d), 82 in (since 8d); 13 
remapped pgs
2023-02-07 22:30:00osd: 82 osds: 82 up (since 9d), 82 in (since 8d); 11 
remapped pgs
2023-02-08 15:30:00osd: 82 osds: 82 up (since 9d), 82 in (since 9d); 9 
remapped pgs
2023-02-08 21:30:00osd: 82 osds: 82 up (since 10d), 82 in (since 9d); 15 remapped 
pgs   <<< increase
2023-02-09 07:30:00osd: 82 osds: 82 up (since 10d), 82 in (since 9d); 13 
remapped pgs
2023-02-09 13:03:27 ceph balancer off
2023-02-09 22:30:00osd: 82 osds: 82

[ceph-users] recovery for node disaster

2023-02-12 Thread farhad kh
 I have a cluster of three nodes, with three replicas per pool on cluster
nodes
-
HOST ADDR LABELS  STATUS
apcepfpspsp0101  192.168.114.157  _admin mon
apcepfpspsp0103  192.168.114.158  mon _admin
apcepfpspsp0105  192.168.114.159  mon _admin
3 hosts in cluster
-
# ceph osd crush rule dump
[
{
"rule_id": 0,
"rule_name": "replicated_rule",
"type": 1,
"steps": [
{
"op": "take",
"item": -1,
"item_name": "default"
},
{
"op": "chooseleaf_firstn",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
}
]
-
epoch 1033
fsid 9c35e594-2392-11ed-809a-005056ae050c
created 2022-08-24T09:53:36.481866+
modified 2023-02-12T18:57:34.447536+
flags sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit
crush_version 51
full_ratio 0.95
backfillfull_ratio 0.9
nearfull_ratio 0.85
require_min_compat_client luminous
min_compat_client luminous
require_osd_release quincy
stretch_mode_enabled false
pool 1 '.mgr' replicated size 3 min_size 2 crush_rule 0 object_hash
rjenkins pg_num 1 pgp_num 1 autoscale_mode on last_change 21 flags
hashpspool stripe_width 0 pg_num_max 32 pg_num_min 1 application mgr
pool 2 'k8s-rbd' replicated size 3 min_size 2 crush_rule 0 object_hash
rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 541 lfor 0/0/44
flags hashpspool,selfmanaged_snaps max_bytes 75161927680 stripe_width 0
application rbd
pool 3 'k8s-cephfs_metadata' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 16 pgp_num 16 autoscale_mode on last_change 543
lfor 0/0/57 flags hashpspool max_bytes 5368709120 stripe_width 0
pg_autoscale_bias 4 pg_num_min 16 recovery_priority 5 application cephfs
pool 4 'k8s-cephfs_data' replicated size 3 min_size 2 crush_rule 0
object_hash rjenkins pg_num 32 pgp_num 32 autoscale_mode on last_change 542
lfor 0/0/57 flags hashpspool max_bytes 32212254720 stripe_width 0
application cephfs
---

Is it possible to recover data when two nodes with all physical disks are
lost for any reason?
What is the maximum number of fault tolerance for the cluster?
For this purpose, consider the default settings .
what changes should I make to increase fault tolerance?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Subject: OSDs added, remapped pgs and objects misplaced cycling up and down

2023-02-12 Thread Alexandre Marangone
This could be the pg autoscaler since you added new OSDs. You can run ceph
osd pool ls detail and check the pg_num and pg_target numbers iirc to
confirm

On Sun, Feb 12, 2023 at 20:24 Chris Dunlop  wrote:

> Hi,
>
> ceph-16.2.9
>
> I've added some new osds - some added to existing hosts and some on
> newly-commissioned hosts. The new osds were added to the data side of an
> existing EC 8.3 pool.
>
> I've been waiting for the system to finish remapping / backfilling for
> some time. Originally the number of remapped pgs and objects misplaced was
> steadily decreasing. However in the last few days both remapped pgs and
> objects misplaced have increased several times. Both then steadily
> decrease before jumping up again. See below
> for some details on this.
>
> I'm guessing these are associated stats, i.e. the number of objects
> misplaced reflects the objects in the remapped pgs so it's not surprising
> they're moving in concert.
>
> I thought at first that this cycling might be due to the balancer (which
> was enabled), but the cycling is continuing after the balancer was
> disabled, e.g.:
>
> $ ts; ceph balancer status
> 2023-02-09 12:39:14
> {
>  "active": true,
>  "last_optimize_duration": "0:00:00.000313",
>  "last_optimize_started": "Thu Feb  9 01:38:55 2023",
>  "mode": "upmap",
>  "optimize_result": "Too many objects (0.078537 > 0.05) are
> misplaced; try again later",
>  "plans": []
> }
> $ ts; ceph balancer off
> 2023-02-09 13:03:27 ceph balancer off
>
> 4 days later, the balancer is still inactive, and "last_optimize_started"
> indicates it hasn't run since before it was turned off:
>
> $ ts; ceph balancer status
> 2023-02-13 10:18:18
> {
>  "active": false,
>  "last_optimize_duration": "0:00:00.000420",
>  "last_optimize_started": "Thu Feb  9 02:02:56 2023",
>  "mode": "upmap",
>  "optimize_result": "Too many objects (0.078101 > 0.05) are
> misplaced; try again later",
>  "plans": []
> }
>
> Since the balancer was turned off the cycling continues: "misplaced" gets
> down to 5% and then rapidly increases to 9%, in conjunction with "starting
> backfill" messages, before starting to slowly decrease once again. See
> below for a log extract.
>
> Is this "sawtooth" pattern of remapped pgs and misplaced objects a normal
> consequence of adding OSDs?
>
> Will it eventually settle done itself or do I need to "do something"?
>
> Is there some way of telling how much work remains until it all settles
> down?
>
> Cheers,
>
> Chris
>
>
> --
> #
> # General health
> #   scrubs disabled until added OSDs complete backfilling etc.
> #
> $ ceph -s
>cluster:
>  id: c6618970-0ce0-4cb2-bc9a-dd5f29b62e24
>  health: HEALTH_WARN
>  noout,noscrub,nodeep-scrub flag(s) set
>  5945 pgs not deep-scrubbed in time
>  5945 pgs not scrubbed in time
>  5 slow ops, oldest one blocked for 464 sec, mon.k2 has slow
> ops
>
>services:
>  mon: 5 daemons, quorum k2,b2,b4,k1,b5 (age 5d)
>  mgr: b4(active, since 3M), standbys: b5, b2
>  osd: 82 osds: 82 up (since 2w), 82 in (since 2w); 12 remapped pgs
>   flags noout,noscrub,nodeep-scrub
>
>data:
>  pools:   16 pools, 5945 pgs
>  objects: 70.35M objects, 88 TiB
>  usage:   186 TiB used, 242 TiB / 428 TiB avail
>  pgs: 50050192/700705500 objects misplaced (7.143%)
>   5933 active+clean
>   10   active+remapped+backfill_wait
>   2active+remapped+backfilling
>
>io:
>  client:   2.2 MiB/s rd, 1.8 MiB/s wr, 395 op/s rd, 122 op/s wr
>  recovery: 24 MiB/s, 24 objects/s
>
> #
> # remapped pgs steadily decreasing then started periodically increasing
> #   (extract from periodic "ceph -s")
> #
> 2023-02-06 00:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d);
> 26 remapped pgs
> 2023-02-06 04:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d);
> 25 remapped pgs
> 2023-02-06 06:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d);
> 24 remapped pgs
> 2023-02-06 09:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 6d);
> 23 remapped pgs
> 2023-02-06 12:30:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d);
> 22 remapped pgs
> 2023-02-06 14:00:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d);
> 21 remapped pgs
> 2023-02-06 14:30:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d);
> 20 remapped pgs
> 2023-02-06 20:30:00osd: 82 osds: 82 up (since 7d), 82 in (since 7d);
> 19 remapped pgs
> 2023-02-07 00:00:00osd: 82 osds: 82 up (since 8d), 82 in (since 7d);
> 18 remapped pgs
> 2023-02-07 01:00:00osd: 82 osds: 82 up (since 8d), 82 in (since 7d);
> 17 remapped pgs
> 2023-02-07 11:00:00osd: 82 osds: 82 up (since 8d), 82 in (since 8d);
> 15 remapped pgs
> 2023-02-07 16:30:00osd: 82 osds: 82 up (since 8d), 82 in (since 8d);
> 13 remapped pgs
> 2023-02-07 22:

[ceph-users] Re: Subject: OSDs added, remapped pgs and objects misplaced cycling up and down

2023-02-12 Thread Chris Dunlop

On Sun, Feb 12, 2023 at 20:24 Chris Dunlop  wrote:

Is this "sawtooth" pattern of remapped pgs and misplaced objects a normal
consequence of adding OSDs?


On Sun, Feb 12, 2023 at 10:02:46PM -0800, Alexandre Marangone wrote:

This could be the pg autoscaler since you added new OSDs. You can run ceph
osd pool ls detail and check the pg_num and pg_target numbers iirc to
confirm


$ ceph osd pool ls detail
... pgp_num 46 pgp_num_target 128 ...

That indeed explains it - thanks!

OK, now I need to find out more about the pg autoscaler:

https://docs.ceph.com/en/latest/rados/operations/placement-groups/

For starters:

$ ceph osd pool autoscale-status | grep -e ^POOL -e ^rbd.ec.data
POOL SIZE  TARGET SIZE   RATE  RAW CAPACITY   RATIO  TARGET RATIO  
EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE  BULK
rbd.ec.data61058G   100.0T  1.375208.0T  0.6610 
 1.0 128  on False

Looks like I maybe should have created that pool with the "--bulk" flag, 
per https://docs.ceph.com/en/latest/rados/operations/placement-groups/


https://docs.ceph.com/en/latest/rados/operations/placement-groups/
--
The autoscaler uses the bulk flag to determine which pool should start 
out with a full complement of PGs and only scales down when the usage 
ratio across the pool is not even. However, if the pool doesn’t have the 
bulk flag, the pool will start out with minimal PGs and only when there is 
more usage in the pool.


I'll wonder if setting "bulk" now will help get stable faster?

$ ceph osd pool set  bulk true

Cheers,

Chris
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io