[ceph-users] Mixed cluster with AMD64 and ARM64 possible?

2025-03-05 Thread Robert Sander

Hi,

just a quick question:

Is a mixed cluster with nodes running AMD64 and ARM64 CPUs possible?

Is the cephadm orchestrator able to manage such a cluster?

Regards
--
Robert Sander
Linux Consultant

Heinlein Consulting GmbH
Schwedter Str. 8/9b, 10119 Berlin

https://www.heinlein-support.de

Tel: +49 30 405051 - 0
Fax: +49 30 405051 - 19

Amtsgericht Berlin-Charlottenburg - HRB 220009 B
Geschäftsführer: Peer Heinlein - Sitz: Berlin
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Mixed cluster with AMD64 and ARM64 possible?

2025-03-05 Thread Adam Prycki

Hi,

we have been running such mixed cluster for almost 2 years.
It works without any cephadm related issues. (Even during upgrade form 
ubuntu 20.04 to 22.04 mixing nodes with docker and podman caused no issues.)


Just be careful when upgrading, ceph arm64 containers for 18.2.4 are 
completely broken. ceph-osd in that containers segvfault and only fix is 
to upgrade to 19.2.0/19.2.1.


https://tracker.ceph.com/issues/67213

Adam

W dniu 6.03.2025 o 08:32, Robert Sander pisze:

Hi,

just a quick question:

Is a mixed cluster with nodes running AMD64 and ARM64 CPUs possible?

Is the cephadm orchestrator able to manage such a cluster?

Regards




smime.p7s
Description: Kryptograficzna sygnatura S/MIME
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] cephfs healthy but mounting it some data cannot be accessed

2025-03-05 Thread xadhoom76
Hi, we have access and we can mount cephfs volume but for some reason data in 
it can be listes but when we try to copy some files from it , some can be and 
some hangs 
Is there a way to fix it or skip the files we cannot access ?
Best regards
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph cluster unable to read/write data properly and cannot recover normally.

2025-03-05 Thread Janne Johansson
Den ons 5 mars 2025 kl 01:59 skrev 小小 <13071007...@163.com>:
> Hi,
>I'm facing a critical issue with my Ceph cluster. It has become unable to 
> read/write data properly and cannot recover normally. What steps should I 
> take to resolve this?

Did you do anything to the cluster, or did anything happen to it or
the network before this situation we see below?

The "6 osds 6 up" makes it look "normal" but have you removed OSDs or
a host that made it end up like this?
It is very strange if OSDs don't go down and their PGs are unknown, so
there is some other information we are lacking here about why so many
PGs are not active even if the OSDs are there.

> osd: 6 osds: 6 up, 6 in; 169 remapped pgs
> pgs: 12.500% pgs unknown
>  6.019% pgs not active
>  27  unknown
>  9   undersized+degraded+peered
>  4   undersized+peered
>   io:
>
> recovery: 1.3 MiB/s, 20 objects/s,

-- 
May the most significant bit of your life be positive.
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Unintuitive (buggy?) CephFS behaviour when dealing with pool_namespace layout attribute

2025-03-05 Thread Florian Haas

Hello everyone,

I'm seeing some behaviour in CephFS that strikes me as unexpected, and I 
wonder if others have thoughts about it.


Consider this scenario:

* Ceph Reef (18.2.4) deployed with Cephadm running on Ubuntu Jammy, 
CephFS client is running kernel 5.15.0-133-generic.
* CephFS is mounted to /mnt, using a CephX identity that has rwp 
permissions.

* CephFS is using a single data pool named cephfs_data.
* In the CephFS root directory, there is a subdirectory named 
"restricted". That subdirectory currently has a single file, 
"test2.bin", of 8 MiB.


There is no other data in the data pool. I can verify this with the 
following rados command:


rados -n client.cephfs -p cephfs_data ls
103.0001
103.

Now I run this, from the root of my mounted CephFS:

cd /mnt
setfattr -n ceph.dir.layout.pool_namespace -v restricted /mnt/restricted

I realise this is naughty, because there is currently a file named 
/mnt/restricted/test2.bin, and I'm not supposed to set layout attributes 
on a non-empty directory. However, the command does succeed, and I am 
able to read back the ceph.dir.layout xattr:


getfattr -n ceph.dir.layout restricted
# file: restricted
ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 
pool=cephfs_data pool_namespace=restricted"


If at this stage I try to read RADOS objects in the "restricted" 
namespace, I get an empty list:


rados -p cephfs_data -N restricted ls
(empty output)

Now I move my test file out of the restricted subdirectory and back in:

mv /mnt/restricted/test2.bin /mnt/
sync
mv /mnt/test2.bin /mnt/restricted/
sync

No change:

rados -p cephfs_data -N restricted ls
(empty output)

Next, I try moving the file out, and *copying* it back in:

mv /mnt/restricted/test2.bin /mnt/
sync
cp  /mnt/test2.bin /mnt/restricted/
sync

Now, I do see objects in the "restricted" namespace:

rados -p cephfs_data -N restricted ls
10001fd.0001
10001fd.

Also, if I create a *new* file in the "restricted" subdirectory, then 
its RADOS objects do end up in the correct namespace:


dd if=/dev/urandom of=/mnt/restricted/test3.bin count=3 bs=4M

rados -p cephfs_data -N restricted ls
10001fe.
10001fe.0002
10001fe.0001
10001fd.0001
10001fd.

By contrast, here's what happens to a non-empty *file,* when I try to 
set ceph.file.layout:


dd if=/dev/urandom of=/mnt/test4.bin count=4 bs=4M

setfattr -n ceph.file.layout.pool_namespace -v restricted test4.bin
setfattr: test4.bin: Directory not empty

So that fails, somewhat expectedly (although the error message is odd).

In summary, there are two things that confuse me here:

1. Why does setting ceph.dir.layout.pool_namespace on a non-empty 
directory succeed, when setting ceph.file.layout.pool_namespace on a 
non-empty file fails (and even confusingly with a "Directory not empty" 
message)?


2. Considering that setting ceph.dir.layout.pool_namespace on a 
non-empty directory does succeed, why does mv'ing a file to a directory 
with a different pool_namespace, and then mv'ing it back, not result in 
its RADOS objects moving to the other namespace?


So I'm curious: is this a bug (or two), or have I been misunderstanding 
what's actually the expected behaviour?


Cheers,
Florian


smime.p7s
Description: S/MIME Cryptographic Signature
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Error Removing Zone from Zonegroup in Multisite Setup

2025-03-05 Thread Mahnoosh Shahidi
Hi everyone,

We have a Ceph multisite setup with a zonegroup containing three zones: one
master and two secondary zones, one of which is designated as an archive.
After recently upgrading to Ceph 19.2.1, I attempted to remove the archive
zone from the zonegroup following the steps in the Ceph documentation.

I ran
radosgw-admin zonegroup remove --rgw-zonegroup=
--rgw-zone= --rgw-realm=
from the cluster containing the archive zone, followed by
radosgw-admin period update --commit --rgw-zonegroup=
--rgw-zone= --rgw-realm=

However, the period update --commit command returns the error:
ERROR: zonegroup  does not contain zone id 
The zone appears correctly in the zonegroup before the removal, and I
expected the archive zone to be successfully removed with the period update
committing without errors.

Is there an additional step required, or could this be a bug related to
Ceph 19.2.1? Any guidance or assistance would be appreciated. Thank you!

Best regards,
Mahnoosh
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Module 'devicehealth' has failed

2025-03-05 Thread Eugen Block
And do you also have the device_health_metrics pool? During one of the  
upgrades to Quincy or so the older device_health_metrics should have  
been renamed. But on one customer cluster I found that both were still  
there, although that didn't cause any trouble. I don't really fully  
grasp yet what exactly goes wrong here, but maybe you have some log  
context as well, not only the stack trace?


Zitat von Alex from North :


yes, I do

.mgr 10 1  769 KiB2  2.3 MiB  04.7 PiB
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Error Removing Zone from Zonegroup in Multisite Setup

2025-03-05 Thread Adam Prycki

Hi,

I've had similar issues before.
radosgw-admin zonegroup/zone tooling is really bad in general.


As far as I know you should remove zone before deleting it form 
zonegroup but I don't think I've ever managed to remove zone cleanly 
from the cluster. (I would be thankful if someone shared a correct way)

(https://www.ibm.com/docs/en/storage-ceph/7?topic=zones-deleting-zone)

To clean this up you may have to dump your zonegroup to file, edit json 
manually and upload it back to ceph without broken zone.


If you have issues with deleting the zone itself from the local cluster 
RGW configuration you may have to hunt down objects containing it's IDs 
in .rgw.root pool. It's messy and dangerous so, don't forget to make a 
backup before you start ;)



best regards,
Adam


W dniu 5.03.2025 o 18:02, Mahnoosh Shahidi pisze:

Hi everyone,

We have a Ceph multisite setup with a zonegroup containing three zones: one
master and two secondary zones, one of which is designated as an archive.
After recently upgrading to Ceph 19.2.1, I attempted to remove the archive
zone from the zonegroup following the steps in the Ceph documentation.

I ran
radosgw-admin zonegroup remove --rgw-zonegroup=
--rgw-zone= --rgw-realm=
from the cluster containing the archive zone, followed by
radosgw-admin period update --commit --rgw-zonegroup=
--rgw-zone= --rgw-realm=

However, the period update --commit command returns the error:
ERROR: zonegroup  does not contain zone id 
The zone appears correctly in the zonegroup before the removal, and I
expected the archive zone to be successfully removed with the period update
committing without errors.

Is there an additional step required, or could this be a bug related to
Ceph 19.2.1? Any guidance or assistance would be appreciated. Thank you!

Best regards,
Mahnoosh
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io




smime.p7s
Description: Kryptograficzna sygnatura S/MIME
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Unintuitive (buggy?) CephFS behaviour when dealing with pool_namespace layout attribute

2025-03-05 Thread Burkhard Linke

Hi,


just some assumptions based on my experience with cephfs:

- you cannot change existing objects; setting a different pool will not 
automagically move data. the data pool seems to be stored in the inode 
information


- this also explains why changing the pool of a file does not work. ceph 
is unable to move the data on its own


- so after changing the pool for a directory, all existing files will 
remain in their original pool


- only newly created files will inherit the changed pool assignment

- moving a file does not create a new inode, so its pool assignment 
cannot be changed by moving



The last time we want to perform a large scale pool change we had to 
copy each file / directory to actually move the data.



Best regards,

Burkhard Linke

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Error Removing Zone from Zonegroup in Multisite Setup

2025-03-05 Thread Shilpa Manjrabad Jagannath
On Wed, Mar 5, 2025 at 9:34 AM Adam Prycki  wrote:

> Hi,
>
> I've had similar issues before.
> radosgw-admin zonegroup/zone tooling is really bad in general.
>
>
> As far as I know you should remove zone before deleting it form
> zonegroup but I don't think I've ever managed to remove zone cleanly
> from the cluster. (I would be thankful if someone shared a correct way)
> (https://www.ibm.com/docs/en/storage-ceph/7?topic=zones-deleting-zone)
>
> To clean this up you may have to dump your zonegroup to file, edit json
> manually and upload it back to ceph without broken zone.
>
> If you have issues with deleting the zone itself from the local cluster
> RGW configuration you may have to hunt down objects containing it's IDs
> in .rgw.root pool. It's messy and dangerous so, don't forget to make a
> backup before you start ;)
>
>
> best regards,
> Adam
>
>
> W dniu 5.03.2025 o 18:02, Mahnoosh Shahidi pisze:
> > Hi everyone,
> >
> > We have a Ceph multisite setup with a zonegroup containing three zones:
> one
> > master and two secondary zones, one of which is designated as an archive.
> > After recently upgrading to Ceph 19.2.1, I attempted to remove the
> archive
> > zone from the zonegroup following the steps in the Ceph documentation.
> >
> > I ran
> > radosgw-admin zonegroup remove --rgw-zonegroup=
> > --rgw-zone= --rgw-realm=
> > from the cluster containing the archive zone, followed by
> > radosgw-admin period update --commit --rgw-zonegroup=
> > --rgw-zone= --rgw-realm=
> >
> > However, the period update --commit command returns the error:
> > ERROR: zonegroup  does not contain zone id 
> > The zone appears correctly in the zonegroup before the removal, and I
> > expected the archive zone to be successfully removed with the period
> update
> > committing without errors.
> >
> > Is there an additional step required, or could this be a bug related to
> > Ceph 19.2.1? Any guidance or assistance would be appreciated. Thank you!
> >
>
>
it does appear to be a regression. have raised a tracker
https://tracker.ceph.com/issues/70323
you don't see that zone entry when you run 'radosgw-admin zonegroup get'
command on the archive zone anymore, do you?

what happens when you run 'radosgw-admin period update' and then
'radosgw-admin period commit' separately?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Unintuitive (buggy?) CephFS behaviour when dealing with pool_namespace layout attribute

2025-03-05 Thread Frédéric Nass
Hi Florian,

Point 1 is certainly a bug regarding the choice of terms in the response 
(confusion between file and directory).

Point 2 is known (cf. https://ewal.dev/cephfs-migrating-files-between-pools) 
and described in the documentation: only new files are written to the new pool 
after setting the new layout. The 'mv' command does not move the Rados objects 
associated with the file from one pool to another.
Only reading and rewriting the file does. What reflections/impediments led to 
this? I do not know.
But what you observed is expected.

Hope this helps,

Cheers,
Frédéric.

De : Florian Haas 
Envoyé : mercredi 5 mars 2025 14:07
À : ceph-users
Objet : [ceph-users] Unintuitive (buggy?) CephFS behaviour when dealing with 
pool_namespace layout attribute

Hello everyone,

I'm seeing some behaviour in CephFS that strikes me as unexpected, and I 
wonder if others have thoughts about it.

Consider this scenario:

* Ceph Reef (18.2.4) deployed with Cephadm running on Ubuntu Jammy, 
CephFS client is running kernel 5.15.0-133-generic.
* CephFS is mounted to /mnt, using a CephX identity that has rwp 
permissions.
* CephFS is using a single data pool named cephfs_data.
* In the CephFS root directory, there is a subdirectory named 
"restricted". That subdirectory currently has a single file, 
"test2.bin", of 8 MiB.

There is no other data in the data pool. I can verify this with the 
following rados command:

rados -n client.cephfs -p cephfs_data ls
103.0001
103.

Now I run this, from the root of my mounted CephFS:

cd /mnt
setfattr -n ceph.dir.layout.pool_namespace -v restricted /mnt/restricted

I realise this is naughty, because there is currently a file named 
/mnt/restricted/test2.bin, and I'm not supposed to set layout attributes 
on a non-empty directory. However, the command does succeed, and I am 
able to read back the ceph.dir.layout xattr:

getfattr -n ceph.dir.layout restricted
# file: restricted
ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304 
pool=cephfs_data pool_namespace=restricted"

If at this stage I try to read RADOS objects in the "restricted" 
namespace, I get an empty list:

rados -p cephfs_data -N restricted ls
(empty output)

Now I move my test file out of the restricted subdirectory and back in:

mv /mnt/restricted/test2.bin /mnt/
sync
mv /mnt/test2.bin /mnt/restricted/
sync

No change:

rados -p cephfs_data -N restricted ls
(empty output)

Next, I try moving the file out, and *copying* it back in:

mv /mnt/restricted/test2.bin /mnt/
sync
cp  /mnt/test2.bin /mnt/restricted/
sync

Now, I do see objects in the "restricted" namespace:

rados -p cephfs_data -N restricted ls
10001fd.0001
10001fd.

Also, if I create a *new* file in the "restricted" subdirectory, then 
its RADOS objects do end up in the correct namespace:

dd if=/dev/urandom of=/mnt/restricted/test3.bin count=3 bs=4M

rados -p cephfs_data -N restricted ls
10001fe.
10001fe.0002
10001fe.0001
10001fd.0001
10001fd.

By contrast, here's what happens to a non-empty *file,* when I try to 
set ceph.file.layout:

dd if=/dev/urandom of=/mnt/test4.bin count=4 bs=4M

setfattr -n ceph.file.layout.pool_namespace -v restricted test4.bin
setfattr: test4.bin: Directory not empty

So that fails, somewhat expectedly (although the error message is odd).

In summary, there are two things that confuse me here:

1. Why does setting ceph.dir.layout.pool_namespace on a non-empty 
directory succeed, when setting ceph.file.layout.pool_namespace on a 
non-empty file fails (and even confusingly with a "Directory not empty" 
message)?

2. Considering that setting ceph.dir.layout.pool_namespace on a 
non-empty directory does succeed, why does mv'ing a file to a directory 
with a different pool_namespace, and then mv'ing it back, not result in 
its RADOS objects moving to the other namespace?

So I'm curious: is this a bug (or two), or have I been misunderstanding 
what's actually the expected behaviour?

Cheers,
Florian
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: ceph orch is not working

2025-03-05 Thread xadhoom76
thanks but it seems another issue. 
Is there any way to upgrade mgr without orchestrator ? or any other services ?
We are online but cannot give any command to cluster
alertmanager.ceph01  ceph01  *:9093,9094  running (10h) 6h ago  21M 
   63.1M-  0.25.0 c8568f914cd2  0bf1d544a775  
ceph-exporter.ceph01 ceph01   running (10h) 6h ago  21M 
   59.0M-  17.2.6 c9a1062f7289  de3f438c448c  
ceph-exporter.ceph02 ceph02   running (10h) 6h ago  21M 
   59.5M-  17.2.6 c9a1062f7289  ba1edb035a66  
crash.ceph01 ceph01   running (7h)  6h ago  21M 
   8640k-  17.2.6 2747c7f13104  e692b2314548  
crash.ceph02 ceph02   running (7h)  6h ago  21M 
   8568k-  17.2.6 2747c7f13104  24535dbbfd76  
crash.ceph03 ceph03   running (7h)  6h ago  21M 
   7394k-  17.2.6 2747c7f13104  86b981a9254c  
grafana.ceph01   ceph01  *:3000   running (10h) 6h ago  21M 
212M-  9.4.7  954c08fa6188  7ef73e324096  
mds.cephfs1.ceph01.niiovaceph01   stopped   6h ago  11h 
   --  
mds.cephfs1.ceph02.qsrlyoceph02   stopped   6h ago  21M 
   --  
mds.cephfs1.ceph03.kuysxqceph03   stopped   6h ago  21M 
   --  
mgr.ceph01.eydqvmceph01  *:8443,9283  running (7h)  6h ago  21M 
439M-  17.2.6 2747c7f13104  6093b54fda30  
mgr.ceph02.wwiqqsceph02  *:8443,9283  running (7h)  6h ago  21M 
576M-  17.2.6 2747c7f13104  bd8a6a0e0dc7  
mgr.ceph03.orwlyvceph03  *:8443,9283  running (7h)  6h ago  21M 
438M-  17.2.6 2747c7f13104  61d563cc5fd2  
mon.ceph01   ceph01   running (7h)  6h ago  21M 
343M2048M  17.2.6 2747c7f13104  33c26b68b7e0  
mon.ceph02   ceph02   running (7h)  6h ago  21M 
324M2048M  17.2.6 2747c7f13104  479224b3c6d9  
mon.ceph03   ceph03   running (7h)  6h ago  21M 
314M2048M  17.2.6 2747c7f13104  affa57e31300  
node-exporter.ceph01 ceph01  *:9100   running (10h) 6h ago  21M 
   37.5M-  1.5.0  0da6a335fe13  a868bdcacffb  
node-exporter.ceph02 ceph02  *:9100   running (10h) 6h ago  21M 
   36.5M-  1.5.0  0da6a335fe13  ffd26fa4c977  
node-exporter.ceph03 ceph03  *:9100   running (10h) 6h ago  21M 
   38.3M-  1.5.0  0da6a335fe13  432fb5b9e903  
osd.0ceph01   running (10h) 6h ago  21M 
   7501M9461M  17.2.8 259b35566514  3a451ff18bbf  
osd.1ceph02   running (10h) 6h ago  21M 
   8623M9589M  17.2.8 259b35566514  a420f4373061  
osd.2ceph03   running (10h) 6h ago   7h 
   3270M9845M  17.2.8 259b35566514  57eba1a3fcaf  
osd.3ceph02   running (10h) 6h ago   7h 
   4580M9589M  17.2.8 259b35566514  6e4ffb81fee5  
osd.4ceph03   running (10h) 6h ago  21M 
   3928M9845M  17.2.8 259b35566514  111b8d5ffb54  
osd.5ceph01   running (10h) 6h ago  21M 
   11.1G9461M  17.2.8 259b35566514  50727e330c7a  
osd.6ceph02   running (10h) 6h ago  21M 
   8811M9589M  17.2.8 259b35566514  602b72c69ab6  
osd.7ceph03   running (10h) 6h ago   7h 
   4491M9845M  17.2.8 259b35566514  1c16697e2c4c  
osd.8ceph01   error 6h ago  21M 
   -9461M  
osd.9ceph02   running (10h) 6h ago  21M 
   9769M9589M  17.2.8 259b35566514  3e812c1dd841  
osd.10   ceph03   running (10h) 6h ago   6h 
   2877M9845M  17.2.8 259b35566514  2102dcc50ead  
osd.11   ceph01   running (10h) 6h ago  21M 
   18.6G9461M  17.2.8 259b35566514  9422bb2e4dca  
osd.12   ceph02   running (10h) 6h ago  21M 
   24.8G9589M  17.2.8 259b35566514  736ae6924f2e  
osd.13   ceph03   running (6h)  6h ago  21M 
   40.0G9845M  17.2.8 259b35566514  b0fd0200dd0e  
osd.14   ceph01   running (10h) 6h ago   6h 
   4578M9461M  17.2.8 259b35566514  ec79e5a13a94  
osd.15   ceph02   running (10h) 6h ago  21M 
   16.7G9589M  17.2.8 259b35566514  a9c876d93119  
osd.16   ceph03

[ceph-users] Re: ceph orch is not working

2025-03-05 Thread Joshua Blanch
I wonder if you are hitting the same issue as this.

https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/D3PHDKKPL7LUHLAMSCO4Y5DESMKIA4FP/

If so, you might have to play around failing the manager so you can stop
the upgrade. AFAIK downgrading versions is not something that is tested at
all.


On Wed, 5 Mar 2025 at 13:03,  wrote:

> Hi , we have issue using ceph orch , it doesn't send any command for real.
> We try to restart or redeploy but nothing works
> This happened downgrading from 17.2.8 to 17.2.6
> Thanks
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>


-- 
Joshua Blanch
Software Engineer

Clyso GmbH
p: +49 89 215252722 | a: Victoria, Canada
w: https://clyso.com | e: joshua.bla...@clyso.com
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: upgrading ceph without orchestrator

2025-03-05 Thread Marc
I would even state updating ceph without the containers is easier.

> 
> Is it possible to upgrade mgr without using ceph orch ?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] ceph orch is not working

2025-03-05 Thread xadhoom76
Hi , we have issue using ceph orch , it doesn't send any command for real.
We try to restart or redeploy but nothing works
This happened downgrading from 17.2.8 to 17.2.6 
Thanks
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Unintuitive (buggy?) CephFS behaviour when dealing with pool_namespace layout attribute

2025-03-05 Thread Gregory Farnum
This is certainly intended behavior. If you checked the layout on the
particular file, you would see it hasn’t changed. Directory layouts are the
default for new files, not a control mechanism for existing files.

It might be confusing, so we can talk about different presentations if
there’s a better option (though this has been the behavior as long as we’ve
had this functionality), but auto-migrating data is a significant challenge
we haven’t sorted through and will have distinct controls if we ever manage
to make it happen.
-Greg

On Wed, Mar 5, 2025 at 5:06 AM Florian Haas  wrote:

> Hello everyone,
>
> I'm seeing some behaviour in CephFS that strikes me as unexpected, and I
> wonder if others have thoughts about it.
>
> Consider this scenario:
>
> * Ceph Reef (18.2.4) deployed with Cephadm running on Ubuntu Jammy,
> CephFS client is running kernel 5.15.0-133-generic.
> * CephFS is mounted to /mnt, using a CephX identity that has rwp
> permissions.
> * CephFS is using a single data pool named cephfs_data.
> * In the CephFS root directory, there is a subdirectory named
> "restricted". That subdirectory currently has a single file,
> "test2.bin", of 8 MiB.
>
> There is no other data in the data pool. I can verify this with the
> following rados command:
>
> rados -n client.cephfs -p cephfs_data ls
> 103.0001
> 103.
>
> Now I run this, from the root of my mounted CephFS:
>
> cd /mnt
> setfattr -n ceph.dir.layout.pool_namespace -v restricted /mnt/restricted
>
> I realise this is naughty, because there is currently a file named
> /mnt/restricted/test2.bin, and I'm not supposed to set layout attributes
> on a non-empty directory. However, the command does succeed, and I am
> able to read back the ceph.dir.layout xattr:
>
> getfattr -n ceph.dir.layout restricted
> # file: restricted
> ceph.dir.layout="stripe_unit=4194304 stripe_count=1 object_size=4194304
> pool=cephfs_data pool_namespace=restricted"
>
> If at this stage I try to read RADOS objects in the "restricted"
> namespace, I get an empty list:
>
> rados -p cephfs_data -N restricted ls
> (empty output)
>
> Now I move my test file out of the restricted subdirectory and back in:
>
> mv /mnt/restricted/test2.bin /mnt/
> sync
> mv /mnt/test2.bin /mnt/restricted/
> sync
>
> No change:
>
> rados -p cephfs_data -N restricted ls
> (empty output)
>
> Next, I try moving the file out, and *copying* it back in:
>
> mv /mnt/restricted/test2.bin /mnt/
> sync
> cp  /mnt/test2.bin /mnt/restricted/
> sync
>
> Now, I do see objects in the "restricted" namespace:
>
> rados -p cephfs_data -N restricted ls
> 10001fd.0001
> 10001fd.
>
> Also, if I create a *new* file in the "restricted" subdirectory, then
> its RADOS objects do end up in the correct namespace:
>
> dd if=/dev/urandom of=/mnt/restricted/test3.bin count=3 bs=4M
>
> rados -p cephfs_data -N restricted ls
> 10001fe.
> 10001fe.0002
> 10001fe.0001
> 10001fd.0001
> 10001fd.
>
> By contrast, here's what happens to a non-empty *file,* when I try to
> set ceph.file.layout:
>
> dd if=/dev/urandom of=/mnt/test4.bin count=4 bs=4M
>
> setfattr -n ceph.file.layout.pool_namespace -v restricted test4.bin
> setfattr: test4.bin: Directory not empty
>
> So that fails, somewhat expectedly (although the error message is odd).
>
> In summary, there are two things that confuse me here:
>
> 1. Why does setting ceph.dir.layout.pool_namespace on a non-empty
> directory succeed, when setting ceph.file.layout.pool_namespace on a
> non-empty file fails (and even confusingly with a "Directory not empty"
> message)?
>
> 2. Considering that setting ceph.dir.layout.pool_namespace on a
> non-empty directory does succeed, why does mv'ing a file to a directory
> with a different pool_namespace, and then mv'ing it back, not result in
> its RADOS objects moving to the other namespace?
>
> So I'm curious: is this a bug (or two), or have I been misunderstanding
> what's actually the expected behaviour?
>
> Cheers,
> Florian
> ___
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] upgrading ceph without orchestrator

2025-03-05 Thread xadhoom76
Is it possible to upgrade mgr without using ceph orch ?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io