[ceph-users] Re: Seeking feedback on Improving cephadm bootstrap process

2023-05-26 Thread Sake Paulusma
Just a user opinion, maybe add the following additions to the options? For option 1: * Clear instructions how to remove all traces to the failed installation (if you can automate it, you can write a manual) or provide instructions to start a cleanup script. * Don't allow another deployment of Ce

[ceph-users] Re: Ceph iscsi gateway semi deprecation warning?

2023-05-26 Thread Alexander E. Patrakov
On Sat, May 27, 2023 at 12:21 AM Mark Kirkwood wrote: > > I am looking at using an iscsi gateway in front of a ceph setup. However > the warning in the docs is concerning: > > The iSCSI gateway is in maintenance as of November 2022. This means that > it is no longer in active development and will

[ceph-users] cephfs-data-scan with multiple data pools

2023-05-26 Thread Justin Li
Dear All, I'm trying to recover failed MDS metadata by following the link below but having troubles. Thanks in advance. Question1: how to scan 2 data pools with scan_extents (cmd 1). The cmd didn't work with two pools specified. Should I scan one then the other? Question2: As to scan_inodes (cmd

[ceph-users] Re: Encryption per user Howto

2023-05-26 Thread Alexander E. Patrakov
On Sat, May 27, 2023 at 5:09 AM Alexander E. Patrakov wrote: > > Hello Frank, > > On Fri, May 26, 2023 at 6:27 PM Frank Schilder wrote: > > > > Hi all, > > > > jumping on this thread as we have requests for which per-client fs mount > > encryption makes a lot of sense: > > > > > What kind of sec

[ceph-users] Re: Encryption per user Howto

2023-05-26 Thread Alexander E. Patrakov
Hello Frank, On Fri, May 26, 2023 at 6:27 PM Frank Schilder wrote: > > Hi all, > > jumping on this thread as we have requests for which per-client fs mount > encryption makes a lot of sense: > > > What kind of security to you want to achieve with encryption keys stored > > on the server side? >

[ceph-users] Re: Seeking feedback on Improving cephadm bootstrap process

2023-05-26 Thread Nico Schottelius
Hello Redouane, much appreciated kick-off for improving cephadm. I was wondering why cephadm does not use a similar approach to rook in the sense of "repeat until it is fixed?" For the background, rook uses a controller that checks the state of the cluster, the state of monitors, whether there

[ceph-users] Important: RGW multisite bug may silently corrupt encrypted objects on replication

2023-05-26 Thread Casey Bodley
Our downstream QE team recently observed an md5 mismatch of replicated objects when testing rgw's server-side encryption in multisite. This corruption is specific to s3 multipart uploads, and only affects the replicated copy - the original object remains intact. The bug likely affects Ceph releases

[ceph-users] Unable to online CephFS, MDS segfaults during mds log replay

2023-05-26 Thread Alfred Heisner
Hello, I have a Ceph deployment using CephFS. Recently MDS failed and cannot start. Attempting to start MDS for this filesystem results in nearly immediate segfault in MDS. Logs below. cephfs-journal-tool shows Overall journal integrity state OK root@proxmox-2:/var/log/ceph# cephfs-journal-too

[ceph-users] Seeking feedback on Improving cephadm bootstrap process

2023-05-26 Thread Redouane Kachach
Dear ceph community, As you are aware, cephadm has become the default tool for installing Ceph on bare-metal systems. Currently, during the bootstrap process of a new cluster, if the user interrupts the process manually or if there are any issues causing the bootstrap process to fail, cephadm leav

[ceph-users] Re: [EXTERNAL] [Pacific] ceph orch device ls do not returns any HDD

2023-05-26 Thread Michel Jouvin
Patrick, I can only say that I would not expect a specific problem due to your hardware. Upgrading the firmware is generally a good idea but I wouldn't expect it helps in your case if the osk (lsblk) sees the disk. As for starting with octopus I don't know if it will help... But we are also

[ceph-users] Re: [EXTERNAL] [Pacific] ceph orch device ls do not returns any HDD

2023-05-26 Thread Patrick Begou
Hi Michel, I do not notice anything strange in the logs files (looking for errors or warnings). The hardware is a DELL C6100 sled (from 2011) running Alma Linux8 up-to-date. It uses 3 sata disks. Is there a way to force osd installation by hand with providing the device /dev/sdc  for examp

[ceph-users] Re: Unexpected behavior of directory mtime after being set explicitly

2023-05-26 Thread Joseph Fernandes
Hello Greogry, We are setting the mtime to 01 Jan 1970 00:00 1. Create a directory "dir1" 2. set mtime of the "dir1 to 0 -> i.e 1 jan 1970 3. Create child directory in "dir1" i.e mkdir dir1/dir2 OR Create a file in "dir1" i.e "touch dir1/file1 4. stat "dir1" Linux FS : updates the

[ceph-users] Re: Unexpected behavior of directory mtime after being set explicitly

2023-05-26 Thread sandip . divekar
Hi Milind It's Kernel Client.  Thanks Sandip ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] sub-read of librados

2023-05-26 Thread johnnyjohnnypd
Hi! I am trying to read only part of an object by specifying the non-trivial offset and length of the read function: `librados::IoCtxImpl::read(const object_t& oid, bufferlist& bl, size_t len, uint64_t off)` from `IoCtxImpl.cc`. However, after connecting to an erasure code pool (e.g., 12+4), I tr

[ceph-users] Re: mgr memory usage constantly increasing

2023-05-26 Thread Simon Fowler
It's a bit of a kludge, but just failing the active mgr on a regular schedule works around this issue (which we see also on our 17.2.5 cluster). We just have a cron job that fails the active mgr every 24 hours - it seems to get up to ~30G, then drop back to 10-15G once it goes to backup mode.

[ceph-users] Ceph iscsi gateway semi deprecation warning?

2023-05-26 Thread Mark Kirkwood
I am looking at using an iscsi gateway in front of a ceph setup. However the warning in the docs is concerning: The iSCSI gateway is in maintenance as of November 2022. This means that it is no longer in active development and will not be updated to add new features. Does this mean I should

[ceph-users] Re: Pacific - MDS behind on trimming

2023-05-26 Thread Dan van der Ster
Hi Emmanuel, In my experience MDS getting behind on trimming normally happens for one of two reasons. Either your client workload is simply too expensive for your metadata pool OSDs to keep up (and btw some ops are known to be quite expensive such as setting xattrs or deleting files). Or I've seen

[ceph-users] Re: [EXTERNAL] [Pacific] ceph orch device ls do not returns any HDD

2023-05-26 Thread Michel Jouvin
Hi Patrick, It is weird, we have a couple of clusters with cephadm and running pacify or quincy and ceph orch device works well. Have you looked at the cephadm logs (ceph log last cephadm)? Except if you are using a very specific hardware, I suspect Ceph is suffering of a problem outside it.

[ceph-users] Re: [EXTERNAL] [Pacific] ceph orch device ls do not returns any HDD

2023-05-26 Thread Patrick Begou
Hi, I'm back working on this problem. First of all, I saw that I had a hardware memory error so I had to solve this first. It's done. I've tested some different Ceph deployments, each time starting with a full OS re-install (it requires some time for each test). Using Octopus, the devices

[ceph-users] Re: mds dump inode crashes file system

2023-05-26 Thread Frank Schilder
Update to the list: a first issue was discovered and fixed on both, the MDS and kclient side. the tracker for the bug is here: https://tracker.ceph.com/issues/61200 . It contains a link to the kclient patchwork. There is no link to the MDS PR (yet). This bug is responsible for the mount going s

[ceph-users] Multi region RGW Config Questions - Quincy

2023-05-26 Thread Deep Dish
Hello, I have a Qunicy (17.2.6) cluster, looking to create a multi-zone / multi-region RGW service and have a few questions with respect to published docs - https://docs.ceph.com/en/quincy/radosgw/multisite/. In general, I understand the process as: 1. Create a new REALM, ZONEGROUP, ZONE

[ceph-users] Re: BlueStore fragmentation woes

2023-05-26 Thread Stefan Kooman
On 5/25/23 22:12, Igor Fedotov wrote: On 25/05/2023 20:36, Stefan Kooman wrote: On 5/25/23 18:17, Igor Fedotov wrote: Perhaps... I don't like the idea to use fragmentation score as a real index. IMO it's mostly like a very imprecise first turn marker to alert that something might be wrong.

[ceph-users] Re: ln: failed to create hard link 'file name': Read-only file system

2023-05-26 Thread Frank Schilder
Update to the list: after extensive debugging by Xiubo on our test cluster, the issue was identified and fixed. A patch in on its way to distro kernels. The tracker for this case is: https://tracker.ceph.com/issues/59515 Best regards, = Frank Schilder AIT Risø Campus Bygning 109,

[ceph-users] Re: Encryption per user Howto

2023-05-26 Thread Frank Schilder
Hi Robert. > But this would still mean that the client encrypts the data. Yes and as far as I understood this would be fine for the original request as well. Maybe this might sound confusing, but here is my terminology for that: I don't count the RGW daemon as a storage server, in my terminolog

[ceph-users] Re: Encryption per user Howto

2023-05-26 Thread Robert Sander
On 5/26/23 12:26, Frank Schilder wrote: It may very well not serve any other purpose, but these are requests we get. If I could provide an encryption key to a ceph-fs kernel at mount time, this requirement could be solved very elegantly on a per-user (request) basis and only making users who

[ceph-users] Re: Encryption per user Howto

2023-05-26 Thread Frank Schilder
Hi all, jumping on this thread as we have requests for which per-client fs mount encryption makes a lot of sense: > What kind of security to you want to achieve with encryption keys stored > on the server side? One of the use cases is if a user requests a share with encryption at rest. Since e

[ceph-users] Re: Help needed to configure erasure coding LRC plugin

2023-05-26 Thread Michel Jouvin
Hi,  I realize that the crushmap I attached to one of my email, probably required to understand the discussion here, has been stripped down by mailman. To avoid poluting the thread with a long output, I put it on at https://box.in2p3.fr/index.php/s/J4fcm7orfNE87CX. Download it if you are inte

[ceph-users] Re: BlueStore fragmentation woes

2023-05-26 Thread Igor Fedotov
yeah, definitely this makes sense On 26/05/2023 09:39, Konstantin Shalygin wrote: Hi Igor, Should we backpot this to the p,q and reef release's? Thanks, k Sent from my iPhone On 25 May 2023, at 23:13, Igor Fedotov wrote: You might be facing the issue fixed by https://github.com/ceph/ceph/

[ceph-users] Re: `ceph features` on Nautilus still reports "luminous"

2023-05-26 Thread Frank Schilder
Hi Oliver, there is a little bit more to it. The feature flags tell you also what version a client should be at a minimum - a bit as indicated by the command name set-require-*min-compat-client*. All clusters allow old clients to connect, but there is a minimum compatibility cap. If you increas

[ceph-users] Pacific - MDS behind on trimming

2023-05-26 Thread Emmanuel Jaep
Hi, lately, we have had some issues with our MDSs (Ceph version 16.2.10 Pacific). Part of them are related to MDS being behind on trimming. I checked the documentation and found the following information ( https://docs.ceph.com/en/pacific/cephfs/health-messages/): > CephFS maintains a metadata j

[ceph-users] Re: Troubleshooting "N slow requests are blocked > 30 secs" on Pacific

2023-05-26 Thread Emmanuel Jaep
Hi Milind, I finally managed to dump the cache and find the file. It generated a 1.5 GB file with about 7 Mio lines. It's kind of hard to know what is out of the ordinary… Furthermore, I noticed that dumping the cache was actually stopping the MDS. Is it a normal behavior? Best, Emmanuel On Th