Just a user opinion, maybe add the following additions to the options?
For option 1:
* Clear instructions how to remove all traces to the failed installation (if
you can automate it, you can write a manual) or provide instructions to start a
cleanup script.
* Don't allow another deployment of Ce
On Sat, May 27, 2023 at 12:21 AM Mark Kirkwood
wrote:
>
> I am looking at using an iscsi gateway in front of a ceph setup. However
> the warning in the docs is concerning:
>
> The iSCSI gateway is in maintenance as of November 2022. This means that
> it is no longer in active development and will
Dear All,
I'm trying to recover failed MDS metadata by following the link below but
having troubles. Thanks in advance.
Question1: how to scan 2 data pools with scan_extents (cmd 1). The cmd
didn't work with two pools specified. Should I scan one then the other?
Question2: As to scan_inodes (cmd
On Sat, May 27, 2023 at 5:09 AM Alexander E. Patrakov
wrote:
>
> Hello Frank,
>
> On Fri, May 26, 2023 at 6:27 PM Frank Schilder wrote:
> >
> > Hi all,
> >
> > jumping on this thread as we have requests for which per-client fs mount
> > encryption makes a lot of sense:
> >
> > > What kind of sec
Hello Frank,
On Fri, May 26, 2023 at 6:27 PM Frank Schilder wrote:
>
> Hi all,
>
> jumping on this thread as we have requests for which per-client fs mount
> encryption makes a lot of sense:
>
> > What kind of security to you want to achieve with encryption keys stored
> > on the server side?
>
Hello Redouane,
much appreciated kick-off for improving cephadm. I was wondering why
cephadm does not use a similar approach to rook in the sense of "repeat
until it is fixed?"
For the background, rook uses a controller that checks the state of the
cluster, the state of monitors, whether there
Our downstream QE team recently observed an md5 mismatch of replicated
objects when testing rgw's server-side encryption in multisite. This
corruption is specific to s3 multipart uploads, and only affects the
replicated copy - the original object remains intact. The bug likely
affects Ceph releases
Hello,
I have a Ceph deployment using CephFS. Recently MDS failed and cannot
start. Attempting to start MDS for this filesystem results in nearly
immediate segfault in MDS. Logs below.
cephfs-journal-tool shows Overall journal integrity state OK
root@proxmox-2:/var/log/ceph# cephfs-journal-too
Dear ceph community,
As you are aware, cephadm has become the default tool for installing Ceph
on bare-metal systems. Currently, during the bootstrap process of a new
cluster, if the user interrupts the process manually or if there are any
issues causing the bootstrap process to fail, cephadm leav
Patrick,
I can only say that I would not expect a specific problem due to your
hardware. Upgrading the firmware is generally a good idea but I wouldn't
expect it helps in your case if the osk (lsblk) sees the disk.
As for starting with octopus I don't know if it will help... But we are
also
Hi Michel,
I do not notice anything strange in the logs files (looking for errors
or warnings).
The hardware is a DELL C6100 sled (from 2011) running Alma Linux8
up-to-date. It uses 3 sata disks.
Is there a way to force osd installation by hand with providing the
device /dev/sdc for examp
Hello Greogry,
We are setting the mtime to 01 Jan 1970 00:00
1. Create a directory "dir1"
2. set mtime of the "dir1 to 0 -> i.e 1 jan 1970
3. Create child directory in "dir1" i.e mkdir dir1/dir2 OR Create a file in
"dir1" i.e "touch dir1/file1
4. stat "dir1"
Linux FS : updates the
Hi Milind
It's Kernel Client.
Thanks
Sandip
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
Hi!
I am trying to read only part of an object by specifying the non-trivial offset
and length of the read function: `librados::IoCtxImpl::read(const object_t&
oid, bufferlist& bl, size_t len, uint64_t off)` from `IoCtxImpl.cc`.
However, after connecting to an erasure code pool (e.g., 12+4), I tr
It's a bit of a kludge, but just failing the active mgr on a regular
schedule works around this issue (which we see also on our 17.2.5
cluster). We just have a cron job that fails the active mgr every 24
hours - it seems to get up to ~30G, then drop back to 10-15G once it
goes to backup mode.
I am looking at using an iscsi gateway in front of a ceph setup. However
the warning in the docs is concerning:
The iSCSI gateway is in maintenance as of November 2022. This means that
it is no longer in active development and will not be updated to add new
features.
Does this mean I should
Hi Emmanuel,
In my experience MDS getting behind on trimming normally happens for
one of two reasons. Either your client workload is simply too
expensive for your metadata pool OSDs to keep up (and btw some ops are
known to be quite expensive such as setting xattrs or deleting files).
Or I've seen
Hi Patrick,
It is weird, we have a couple of clusters with cephadm and running pacify
or quincy and ceph orch device works well. Have you looked at the cephadm
logs (ceph log last cephadm)?
Except if you are using a very specific hardware, I suspect Ceph is
suffering of a problem outside it.
Hi,
I'm back working on this problem.
First of all, I saw that I had a hardware memory error so I had to solve
this first. It's done.
I've tested some different Ceph deployments, each time starting with a
full OS re-install (it requires some time for each test).
Using Octopus, the devices
Update to the list: a first issue was discovered and fixed on both, the MDS and
kclient side. the tracker for the bug is here:
https://tracker.ceph.com/issues/61200 . It contains a link to the kclient
patchwork. There is no link to the MDS PR (yet).
This bug is responsible for the mount going s
Hello,
I have a Qunicy (17.2.6) cluster, looking to create a multi-zone /
multi-region RGW service and have a few questions with respect to published
docs - https://docs.ceph.com/en/quincy/radosgw/multisite/.
In general, I understand the process as:
1. Create a new REALM, ZONEGROUP, ZONE
On 5/25/23 22:12, Igor Fedotov wrote:
On 25/05/2023 20:36, Stefan Kooman wrote:
On 5/25/23 18:17, Igor Fedotov wrote:
Perhaps...
I don't like the idea to use fragmentation score as a real index. IMO
it's mostly like a very imprecise first turn marker to alert that
something might be wrong.
Update to the list: after extensive debugging by Xiubo on our test cluster, the
issue was identified and fixed. A patch in on its way to distro kernels. The
tracker for this case is: https://tracker.ceph.com/issues/59515
Best regards,
=
Frank Schilder
AIT Risø Campus
Bygning 109,
Hi Robert.
> But this would still mean that the client encrypts the data.
Yes and as far as I understood this would be fine for the original request as
well. Maybe this might sound confusing, but here is my terminology for that:
I don't count the RGW daemon as a storage server, in my terminolog
On 5/26/23 12:26, Frank Schilder wrote:
It may very well not serve any other purpose, but these are requests we get. If
I could provide an encryption key to a ceph-fs kernel at mount time, this
requirement could be solved very elegantly on a per-user (request) basis and
only making users who
Hi all,
jumping on this thread as we have requests for which per-client fs mount
encryption makes a lot of sense:
> What kind of security to you want to achieve with encryption keys stored
> on the server side?
One of the use cases is if a user requests a share with encryption at rest.
Since e
Hi,
I realize that the crushmap I attached to one of my email, probably
required to understand the discussion here, has been stripped down by
mailman. To avoid poluting the thread with a long output, I put it on at
https://box.in2p3.fr/index.php/s/J4fcm7orfNE87CX. Download it if you are
inte
yeah, definitely this makes sense
On 26/05/2023 09:39, Konstantin Shalygin wrote:
Hi Igor,
Should we backpot this to the p,q and reef release's?
Thanks,
k
Sent from my iPhone
On 25 May 2023, at 23:13, Igor Fedotov wrote:
You might be facing the issue fixed by https://github.com/ceph/ceph/
Hi Oliver,
there is a little bit more to it. The feature flags tell you also what version
a client should be at a minimum - a bit as indicated by the command name
set-require-*min-compat-client*. All clusters allow old clients to connect, but
there is a minimum compatibility cap. If you increas
Hi,
lately, we have had some issues with our MDSs (Ceph version 16.2.10
Pacific).
Part of them are related to MDS being behind on trimming.
I checked the documentation and found the following information (
https://docs.ceph.com/en/pacific/cephfs/health-messages/):
> CephFS maintains a metadata j
Hi Milind,
I finally managed to dump the cache and find the file.
It generated a 1.5 GB file with about 7 Mio lines. It's kind of hard to
know what is out of the ordinary…
Furthermore, I noticed that dumping the cache was actually stopping the
MDS. Is it a normal behavior?
Best,
Emmanuel
On Th
31 matches
Mail list logo