That kind of information is ALWAYS in the headers of every email.

> List-Unsubscribe: <mailto:ceph-users-le...@ceph.io>

> On Jun 25, 2020, at 9:21 AM, adan <lyf...@gmail.com> wrote:
> 
> hello
> 
> i want to unsubscribe this mail list . help me please.
> 
> 
> 在 2020/6/25 22:42, ceph-users-requ...@ceph.io 写道:
>> Send ceph-users mailing list submissions to
>>      ceph-users@ceph.io
>> 
>> To subscribe or unsubscribe via email, send a message with subject or
>> body 'help' to
>>      ceph-users-requ...@ceph.io
>> 
>> You can reach the person managing the list at
>>      ceph-users-ow...@ceph.io
>> 
>> When replying, please edit your Subject line so it is more specific
>> than "Re: Contents of ceph-users digest..."
>> 
>> Today's Topics:
>> 
>>    1. Re: Removing pool in nautilus is incredibly slow
>>       (Francois Legrand)
>>    2. Re: Bench on specific OSD (Marc Roos)
>>    3. Re: Removing pool in nautilus is incredibly slow
>>       (Wout van Heeswijk)
>>    4. Lifecycle message on logs (Marcelo Miziara)
>>    5. Re: Feedback of the used configuration (Simon Sutter)
>>    6. Re: Removing pool in nautilus is incredibly slow
>>       (Francois Legrand)
>>    7. Re: Removing pool in nautilus is incredibly slow (Eugen Block)
>> 
>> 
>> ----------------------------------------------------------------------
>> 
>> Date: Thu, 25 Jun 2020 07:57:48 -0000
>> From: "Francois Legrand" <f...@lpnhe.in2p3.fr>
>> Subject: [ceph-users] Re: Removing pool in nautilus is incredibly slow
>> To: ceph-users@ceph.io
>> Message-ID: <159307186843.20.7354239610800797635@mailman-web>
>> Content-Type: text/plain; charset="utf-8"
>> 
>> Does someone have an idea ?
>> F.
>> 
>> ------------------------------
>> 
>> Date: Thu, 25 Jun 2020 11:36:24 +0200
>> From: "Marc Roos" <m.r...@f1-outsourcing.eu>
>> Subject: [ceph-users] Re: Bench on specific OSD
>> To: ceph-users <ceph-users@ceph.io>, seenafallah
>>      <seenafal...@gmail.com>
>> Message-ID: <"H0000071001729d1.1593077784.sx.f1-outsourcing.eu*"@MHS>
>> Content-Type: text/plain;       charset="US-ASCII"
>> 
>>  
>> What is wrong with just doing multiple tests and group in your charts
>> osd's by host?
>> 
>> 
>> -----Original Message-----
>> To: ceph-users
>> Subject: [ceph-users] Bench on specific OSD
>> 
>> Hi all.
>> 
>> Is there anyway to completely health check one OSD host or instance?
>> For example rados bech just on that OSD or do some checks for disk and
>> front and back netowrk?
>> 
>> Thanks.
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
>> email to ceph-users-le...@ceph.io
>> 
>> 
>> ------------------------------
>> 
>> Date: Thu, 25 Jun 2020 14:26:59 +0200
>> From: Wout van Heeswijk <w...@42on.com>
>> Subject: [ceph-users] Re: Removing pool in nautilus is incredibly slow
>> To: f...@lpnhe.in2p3.fr
>> Cc: ceph-users@ceph.io
>> Message-ID: <38a5d557-7797-a68a-1b67-4c8f0e1ec...@42on.com>
>> Content-Type: text/plain; charset=utf-8; format=flowed
>> 
>> Hi Francois,
>> 
>> Have you already looked at the option "osd_delete_sleep"? It will not
>> speed up the process but I will give you some control over your cluster
>> performance.
>> 
>> Something like:
>> 
>> ceph tell osd.\* injectargs '--osd_delete_sleep1'
>> 
>> kind regards,
>> 
>> Wout
>> 42on
>> 
>> On 25-06-2020 09:57, Francois Legrand wrote:
>>> Does someone have an idea ?
>>> F.
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@ceph.io
>>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
>> ------------------------------
>> 
>> Date: Thu, 25 Jun 2020 09:53:09 -0300
>> From: Marcelo Miziara <raxi...@gmail.com>
>> Subject: [ceph-users] Lifecycle message on logs
>> To: ceph-users@ceph.io
>> Message-ID:
>>      <CAGxsqB296eczKux19OKObkimXg3PZ_UDEWVTTqEnrAf=agm...@mail.gmail.com>
>> Content-Type: text/plain; charset="UTF-8"
>> 
>> Hello...it's the first time I need to use the lifecycle, and I created a
>> bucket and set it to expire in one day with s3cmd:
>>  s3cmd expire --expiry-days=1 s3://bucket
>> 
>> The rgw_lifecycle_work_time is set to the default values(00:00-06:00). But
>> I noticed in the rgw logs a lot of messages like:
>> 2020-06-16 00:00:00.311369 7fe2cac87700  0 RGWLC::process() failed to get
>> obj entry lc.8
>> 2020-06-16 00:00:00.311623 7fe2c8c83700  0 RGWLC::process() failed to get
>> obj entry lc.16
>> 2020-06-16 00:00:00.311862 7fe2c6c7f700  0 RGWLC::process() failed to get
>> obj entry lc.4
>> 2020-06-16 00:00:00.319424 7fe2cac87700  0 RGWLC::process() failed to get
>> obj entry lc.10
>> 2020-06-16 00:00:00.319647 7fe2c8c83700  0 RGWLC::process() failed to get
>> obj entry lc.18
>> 2020-06-16 00:00:00.320682 7fe2c6c7f700  0 RGWLC::process() failed to get
>> obj entry lc.16
>> 2020-06-16 00:00:00.327770 7fe2cac87700  0 RGWLC::process() failed to get
>> obj entry lc.6
>> 2020-06-16 00:00:00.328941 7fe2c8c83700  0 RGWLC::process() failed to get
>> obj entry lc.17
>> 2020-06-16 00:00:00.332463 7fe2c6c7f700  0 RGWLC::process() failed to get
>> obj entry lc.20
>> 2020-06-16 00:00:00.336788 7fe2cac87700  0 RGWLC::process() failed to get
>> obj entry lc.1
>> 2020-06-16 00:00:00.336924 7fe2c8c83700  0 RGWLC::process() failed to get
>> obj entry lc.24
>> 2020-06-16 00:00:00.340915 7fe2c6c7f700  0 RGWLC::process() failed to get
>> obj entry lc.2
>> 
>> The object was deleted, but these messages keep appearing.
>> Is it safe to ignore them?
>> 
>> For the records, i'm using redhat luminous 12.2.12
>> 
>> Thanks, Marcelo.
>> 
>> ------------------------------
>> 
>> Date: Thu, 25 Jun 2020 13:19:38 +0000
>> From: Simon Sutter <ssut...@hosttech.ch>
>> Subject: [ceph-users] Re: Feedback of the used configuration
>> To: Paul Emmerich <paul.emmer...@croit.io>
>> Cc: "ceph-users@ceph.io" <ceph-users@ceph.io>
>> Message-ID: <f6e11ee189a74c35815c495d64cbe...@hosttech.ch>
>> Content-Type: text/plain; charset="utf-8"
>> 
>> Hello Paul,
>> 
>> Thanks for the Answer.
>> I took a look at the subvolumes, but they are a bit odd in my opinion.
>> If I create one with a subvolume-group, the folder structure will look like 
>> this:
>> /cephfs/volumes/group-name/subvolume-name/random-uuid/
>> And I have to issue two commands, first set the group and then set the 
>> volume name, but why so complicated?
>> 
>> Wouldn't it be easier to  just make subvolumes anywhere inside the cephfs?
>> I can see the intended use for groups, but if I want to publish a pool in 
>> some different directory that's not possible (except for setfattr).
>> Without first creating subvolume-groups, the orchestrator creates subvolumes 
>> in the /cephfs/volumes/_nogroup/subvolume-name/randmon-uuid/ folder.
>> 
>> And the more important question is, why is there a new folder with a random 
>> uuid inside the subvolume?
>> I try to understand the points the devs had, when they developed this, but 
>> for me, this is something I have to explain to some devs in our team and at 
>> the moment I can't.
>> 
>> It is indeed easier to deploy but comes with much less flexibility.
>> Maybe something to write in the tracker about?
>> 
>> Thanks in advance,
>> Simon
>> 
>> Von: Paul Emmerich [mailto:paul.emmer...@croit.io]
>> Gesendet: Mittwoch, 24. Juni 2020 17:35
>> An: Simon Sutter <ssut...@hosttech.ch>
>> Cc: ceph-users@ceph.io
>> Betreff: Re: [ceph-users] Feedback of the used configuration
>> 
>> Have a look at cephfs subvolumes: 
>> https://docs.ceph.com/docs/master/cephfs/fs-volumes/#fs-subvolumes
>> 
>> They are internally just directories with quota/pool placement 
>> layout/namespace with some mgr magic to make it easier than doing that all 
>> by hand
>> 
>> Paul
>> 
>> --
>> Paul Emmerich
>> 
>> Looking for help with your Ceph cluster? Contact us at https://croit.io
>> 
>> croit GmbH
>> Freseniusstr. 31h
>> 81247 München
>> www.croit.io<http://www.croit.io>
>> Tel: +49 89 1896585 90
>> 
>> 
>> On Wed, Jun 24, 2020 at 4:38 PM Simon Sutter 
>> <ssut...@hosttech.ch<mailto:ssut...@hosttech.ch>> wrote:
>> Hello,
>> 
>> After two months of the "ceph try and error game", I finally managed to get 
>> an Octopuss cluster up and running.
>> The unconventional thing about it is, it's just for hot backups, no virtual 
>> machines on there.
>> All the  nodes are without any caching ssd's, just plain hdd's.
>> At the moment there are eight of them with a total of 50TB. We are planning 
>> to go up to 25 and bigger disks so we end on 300TB-400TB
>> 
>> I decided to go with cephfs, because I don't have any experience in things 
>> like S3 and I need to read the same file system from more than one client.
>> 
>> I made one cephfs with a replicated pool.
>> On there I added erasure-coded pools to save some Storage.
>> To add those pools, I did it with the setfattr command like this:
>> setfattr -n ceph.dir.layout.pool -v ec_data_server1 /cephfs/nfs/server1
>> 
>> Some of our servers cannot use cephfs (old kernels, special OS's) so I have 
>> to use nfs.
>> This is set up with the included ganesha-nfs.
>> Exported is the /cephfs/nfs folder and clients can mount folders below this.
>> 
>> There are two final questions:
>> 
>> -          Was it right to go with the way of "mounting" pools with 
>> setfattr, or should I have used multiple cephfs?
>> 
>> First I was thinking about using multiple cephfs but there are warnings 
>> everywhere. The deeper I got in, the more it seems I would have been fine 
>> with multiple cephfs.
>> 
>> -          Is there a way I don't know, but it would be easier?
>> 
>> I still don't know much about Rest, S3, RBD etc... so there may be a better 
>> way
>> 
>> Other remarks are desired.
>> 
>> Thanks in advance,
>> Simon
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io<mailto:ceph-users@ceph.io>
>> To unsubscribe send an email to 
>> ceph-users-le...@ceph.io<mailto:ceph-users-le...@ceph.io>
>> 
>> ------------------------------
>> 
>> Date: Thu, 25 Jun 2020 16:22:12 +0200
>> From: Francois Legrand <f...@lpnhe.in2p3.fr>
>> Subject: [ceph-users] Re: Removing pool in nautilus is incredibly slow
>> To: Wout van Heeswijk <w...@42on.com>
>> Cc: ceph-users@ceph.io
>> Message-ID: <db34e791-1172-8b94-09e7-4ab5790d3...@lpnhe.in2p3.fr>
>> Content-Type: text/plain; charset=utf-8; format=flowed
>> 
>> Thanks for the hint.
>> I tryed but it doesn't seems to change anything...
>> Moreover, as the osds seems quite loaded I had regularly some osd marked
>> down which triggered some new peering and thus more load !!!
>> I set the osd no down flag, but I still have some osd reported (wrongly)
>> as down (and back up in the minute) which generate peering and
>> remapping. I don't really understand the action of no down parameter !
>> Is there a way to tell ceph not to peer immediately after an osd is
>> reported down (let say wait for 60s) ?
>> I am thinking about restarting all osd (or maybe the whole cluster) to
>> get osd_op_queue_cut_off changed to high and osd_op_thread_timeout to
>> something higher than 15 (but I don't think it will really improve the
>> situation).
>> F.
>> 
>> 
>> Le 25/06/2020 à 14:26, Wout van Heeswijk a écrit :
>>> Hi Francois,
>>> 
>>> Have you already looked at the option "osd_delete_sleep"? It will not
>>> speed up the process but I will give you some control over your
>>> cluster performance.
>>> 
>>> Something like:
>>> 
>>> ceph tell osd.\* injectargs '--osd_delete_sleep1'
>>> kind regards,
>>> 
>>> Wout
>>> 42on
>>> On 25-06-2020 09:57, Francois Legrand wrote:
>>>> Does someone have an idea ?
>>>> F.
>>>> _______________________________________________
>>>> ceph-users mailing list --ceph-users@ceph.io
>>>> To unsubscribe send an email toceph-users-le...@ceph.io
>> 
>> ------------------------------
>> 
>> Date: Thu, 25 Jun 2020 14:42:57 +0000
>> From: Eugen Block <ebl...@nde.ag>
>> Subject: [ceph-users] Re: Removing pool in nautilus is incredibly slow
>> To: ceph-users@ceph.io
>> Message-ID:
>>      <20200625144257.horde.i8jpcddeor47wdoehqkq...@webmail.nde.ag>
>> Content-Type: text/plain; charset=utf-8; format=flowed; DelSp=Yes
>> 
>> I'm not sure if your OSDs have their rocksDB on faster devices, if not
>> it sounds a lot like rocksdb fragmentation [1] leading to a very high
>> load on the OSDs and occasionally crashing OSDs. If you don't plan to
>> delete so much data at once on a regular basis you could sit this one
>> out, but one solution is to re-create the OSDs with rocksDB/WAL on
>> faster devices.
>> 
>> 
>> [1] https://www.mail-archive.com/ceph-users@ceph.io/msg03160.html
>> 
>> 
>> Zitat von Francois Legrand <f...@lpnhe.in2p3.fr>:
>> 
>>> Thanks for the hint.
>>> I tryed but it doesn't seems to change anything...
>>> Moreover, as the osds seems quite loaded I had regularly some osd
>>> marked down which triggered some new peering and thus more load !!!
>>> I set the osd no down flag, but I still have some osd reported
>>> (wrongly) as down (and back up in the minute) which generate peering
>>> and remapping. I don't really understand the action of no down
>>> parameter !
>>> Is there a way to tell ceph not to peer immediately after an osd is
>>> reported down (let say wait for 60s) ?
>>> I am thinking about restarting all osd (or maybe the whole cluster)
>>> to get osd_op_queue_cut_off changed to high and
>>> osd_op_thread_timeout to something higher than 15 (but I don't think
>>> it will really improve the situation).
>>> F.
>>> 
>>> 
>>> Le 25/06/2020 à 14:26, Wout van Heeswijk a écrit :
>>>> Hi Francois,
>>>> 
>>>> Have you already looked at the option "osd_delete_sleep"? It will
>>>> not speed up the process but I will give you some control over your
>>>> cluster performance.
>>>> 
>>>> Something like:
>>>> 
>>>> ceph tell osd.\* injectargs '--osd_delete_sleep1'
>>>> kind regards,
>>>> 
>>>> Wout
>>>> 42on
>>>> On 25-06-2020 09:57, Francois Legrand wrote:
>>>>> Does someone have an idea ?
>>>>> F.
>>>>> _______________________________________________
>>>>> ceph-users mailing list --ceph-users@ceph.io
>>>>> To unsubscribe send an email toceph-users-le...@ceph.io
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@ceph.io
>>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> 
>> 
>> ------------------------------
>> 
>> Subject: Digest Footer
>> 
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@ceph.io
>> To unsubscribe send an email to ceph-users-le...@ceph.io
>> %(web_page_url)slistinfo%(cgiext)s/%(_internal_name)s
>> 
>> 
>> ------------------------------
>> 
>> End of ceph-users Digest, Vol 89, Issue 112
>> *******************************************
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to