Hello all:
1 I want to use mClockClientQueue to restrict background operations
like scrub and recovery,but official documents say this is still in
the experimental stage, so I would like to ask if there are any
problems with actual use.
2 I want to subscribe ceph-devel group and I have sent an email
"subscribe ceph devel " to ceph-de...@vger.kernel.org but cannot join
in. How can I join?


                                                                  From
      WeiHaocheng
<ceph-users-requ...@lists.ceph.com> 于2018年9月30日周日 上午7:39写道:
>
> Send ceph-users mailing list submissions to
>         ceph-users@lists.ceph.com
>
> To subscribe or unsubscribe via the World Wide Web, visit
>         http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> or, via email, send a message with subject or body 'help' to
>         ceph-users-requ...@lists.ceph.com
>
> You can reach the person managing the list at
>         ceph-users-ow...@lists.ceph.com
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of ceph-users digest..."
>
>
> Today's Topics:
>
>    1. Re: Manually deleting an RGW bucket (Konstantin Shalygin)
>    2. mount cephfs from a public network ip of mds (Joshua Chen)
>    3. Re: mount cephfs from a public network ip of mds (Paul Emmerich)
>    4. Re: Any backfill in our cluster makes the cluster unusable
>       and takes forever (Pavan Rallabhandi)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Sat, 29 Sep 2018 13:39:32 +0700
> From: Konstantin Shalygin <k0...@k0ste.ru>
> To: ceph-users@lists.ceph.com
> Subject: Re: [ceph-users] Manually deleting an RGW bucket
> Message-ID: <1738d937-1cdd-da2d-11ed-0f2429cbb...@k0ste.ru>
> Content-Type: text/plain; charset=utf-8; format=flowed
>
> > How do I delete an RGW/S3 bucket and its contents if the usual S3 API 
> > commands don't work?
> >
> > The bucket has S3 delete markers that S3 API commands are not able to 
> > remove, and I'd like to reuse the bucket name.  It was set up for 
> > versioning and lifecycles under ceph 12.2.5 which broke the bucket when a 
> > reshard happened.  12.2.7 allowed me to remove the regular files but not 
> > the delete markers.
> >
> > There must be a way of removing index files and so forth through rados 
> > commands.
>
>
> What error actually is?
>
> For delete bucket you should delete all bucket objects ("s3cmd rm -rf
> s3://bucket/") and multipart uploads.
>
>
>
> k
>
>
>
> ------------------------------
>
> Message: 2
> Date: Sat, 29 Sep 2018 18:07:20 +0800
> From: Joshua Chen <csc...@asiaa.sinica.edu.tw>
> To: ceph-users <ceph-users@lists.ceph.com>
> Subject: [ceph-users] mount cephfs from a public network ip of mds
> Message-ID:
>         <CAOUXHtg9sTX9ex1YPTcHpP=aux2sph2y3efbwq-bzwu_40x...@mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> Hello all,
>   I am testing the cephFS cluster so that clients could mount -t ceph.
>
>   the cluster has 6 nodes, 3 mons (also mds), and 3 osds.
>   All these 6 nodes has 2 nic, one 1Gb nic with real ip (140.109.0.0) and 1
> 10Gb nic with virtual ip (10.32.0.0)
>
> 140.109. Nic1 1G<-MDS1->Nic2 10G 10.32.
> 140.109. Nic1 1G<-MDS2->Nic2 10G 10.32.
> 140.109. Nic1 1G<-MDS3->Nic2 10G 10.32.
> 140.109. Nic1 1G<-OSD1->Nic2 10G 10.32.
> 140.109. Nic1 1G<-OSD2->Nic2 10G 10.32.
> 140.109. Nic1 1G<-OSD3->Nic2 10G 10.32.
>
>
>
> and I have the following questions:
>
> 1, can I have both public (140.109.0.0) and cluster (10.32.0.0) clients all
> be able to mount this cephfs resource
>
> I want to do
>
> (in a 140.109 network client)
> mount -t ceph mds1(140.109.169.48):/ /mnt/cephfs -o user=,secret=,,,,
>
> and also in a 10.32.0.0 network client)
> mount -t ceph mds1(10.32.67.48):/
> /mnt/cephfs -o user=,secret=,,,,
>
>
>
>
> Currently, only this 10.32.0.0 clients can mount it. that of public network
> (140.109) can not. How can I enable this?
>
> here attached is my ceph.conf
>
> Thanks in advance
>
> Cheers
> Joshua
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: 
> <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20180929/aad45a46/attachment-0001.html>
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: ceph.conf
> Type: application/octet-stream
> Size: 304 bytes
> Desc: not available
> URL: 
> <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20180929/aad45a46/attachment-0001.obj>
>
> ------------------------------
>
> Message: 3
> Date: Sat, 29 Sep 2018 12:42:36 +0200
> From: Paul Emmerich <paul.emmer...@croit.io>
> To: Joshua Chen <csc...@asiaa.sinica.edu.tw>
> Cc: Ceph Users <ceph-users@lists.ceph.com>
> Subject: Re: [ceph-users] mount cephfs from a public network ip of mds
> Message-ID:
>         <CAD9yTbEEtFSHFjDp7NteMS9pJfjEiB_8+grC-Y=urndfju-...@mail.gmail.com>
> Content-Type: text/plain; charset="UTF-8"
>
> All Ceph clients will always first connect to the mons. Mons provide
> further information on the cluster such as the IPs of MDS and OSDs.
>
> This means you need to provide the mon IPs to the mount command, not
> the MDS IPs. Your first command works by coincidence since
> you seem to run the mons and MDS' on the same server.
>
>
> Paul
> Am Sa., 29. Sep. 2018 um 12:07 Uhr schrieb Joshua Chen
> <csc...@asiaa.sinica.edu.tw>:
> >
> > Hello all,
> >   I am testing the cephFS cluster so that clients could mount -t ceph.
> >
> >   the cluster has 6 nodes, 3 mons (also mds), and 3 osds.
> >   All these 6 nodes has 2 nic, one 1Gb nic with real ip (140.109.0.0) and 1 
> > 10Gb nic with virtual ip (10.32.0.0)
> >
> > 140.109. Nic1 1G<-MDS1->Nic2 10G 10.32.
> > 140.109. Nic1 1G<-MDS2->Nic2 10G 10.32.
> > 140.109. Nic1 1G<-MDS3->Nic2 10G 10.32.
> > 140.109. Nic1 1G<-OSD1->Nic2 10G 10.32.
> > 140.109. Nic1 1G<-OSD2->Nic2 10G 10.32.
> > 140.109. Nic1 1G<-OSD3->Nic2 10G 10.32.
> >
> >
> >
> > and I have the following questions:
> >
> > 1, can I have both public (140.109.0.0) and cluster (10.32.0.0) clients all 
> > be able to mount this cephfs resource
> >
> > I want to do
> >
> > (in a 140.109 network client)
> > mount -t ceph mds1(140.109.169.48):/ /mnt/cephfs -o user=,secret=,,,,
> >
> > and also in a 10.32.0.0 network client)
> > mount -t ceph mds1(10.32.67.48):/
> > /mnt/cephfs -o user=,secret=,,,,
> >
> >
> >
> >
> > Currently, only this 10.32.0.0 clients can mount it. that of public network 
> > (140.109) can not. How can I enable this?
> >
> > here attached is my ceph.conf
> >
> > Thanks in advance
> >
> > Cheers
> > Joshua
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> --
> Paul Emmerich
>
> Looking for help with your Ceph cluster? Contact us at https://croit.io
>
> croit GmbH
> Freseniusstr. 31h
> 81247 M?nchen
> www.croit.io
> Tel: +49 89 1896585 90
>
>
> ------------------------------
>
> Message: 4
> Date: Sat, 29 Sep 2018 17:57:12 +0000
> From: Pavan Rallabhandi <prallabha...@walmartlabs.com>
> To: David Turner <drakonst...@gmail.com>
> Cc: ceph-users <ceph-users@lists.ceph.com>
> Subject: Re: [ceph-users] Any backfill in our cluster makes the
>         cluster unusable and takes forever
> Message-ID: <18af3764-c26c-4003-b8e6-823ed4bae...@walmartlabs.com>
> Content-Type: text/plain; charset="utf-8"
>
> I looked at one of my test clusters running Jewel on Ubuntu 16.04, and 
> interestingly I found this(below) in one of the OSD logs, which is different 
> from your OSD boot log, where none of the compression algorithms seem to be 
> supported. This hints more at how rocksdb was built on CentOS for Ceph.
>
> 2018-09-29 17:38:38.629112 7fbd318d4b00  4 rocksdb: Compression algorithms 
> supported:
> 2018-09-29 17:38:38.629112 7fbd318d4b00  4 rocksdb:     Snappy supported: 1
> 2018-09-29 17:38:38.629113 7fbd318d4b00  4 rocksdb:     Zlib supported: 1
> 2018-09-29 17:38:38.629113 7fbd318d4b00  4 rocksdb:     Bzip supported: 0
> 2018-09-29 17:38:38.629114 7fbd318d4b00  4 rocksdb:     LZ4 supported: 0
> 2018-09-29 17:38:38.629114 7fbd318d4b00  4 rocksdb:     ZSTD supported: 0
> 2018-09-29 17:38:38.629115 7fbd318d4b00  4 rocksdb: Fast CRC32 supported: 0
>
> ?On 9/27/18, 2:56 PM, "Pavan Rallabhandi" <prallabha...@walmartlabs.com> 
> wrote:
>
>     I see Filestore symbols on the stack, so the bluestore config doesn?t 
> affect. And the top frame of the stack hints at a RocksDB issue, and there 
> are a whole lot of these too:
>
>     ?2018-09-17 19:23:06.480258 7f1f3d2a7700  2 rocksdb: 
> [/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.2.4/rpm/el7/BUILD/ceph-12.2.4/src/rocksdb/table/block_based_table_reader.cc:636]
>  Cannot find Properties block from file.?
>
>     It really seems to be something with RocksDB on centOS. I still think you 
> can try removing ?compression=kNoCompression? from the 
> filestore_rocksdb_options And/Or check if rocksdb is expecting snappy to be 
> enabled.
>
>     Thanks,
>     -Pavan.
>
>     From: David Turner <drakonst...@gmail.com>
>     Date: Thursday, September 27, 2018 at 1:18 PM
>     To: Pavan Rallabhandi <prallabha...@walmartlabs.com>
>     Cc: ceph-users <ceph-users@lists.ceph.com>
>     Subject: EXT: Re: [ceph-users] Any backfill in our cluster makes the 
> cluster unusable and takes forever
>
>     I got pulled away from this for a while.  The error in the log is "abort: 
> Corruption: Snappy not supported or corrupted Snappy compressed block 
> contents" and the OSD has 2 settings set to snappy by default, 
> async_compressor_type and bluestore_compression_algorithm.  Do either of 
> these settings affect the omap store?
>
>     On Wed, Sep 19, 2018 at 2:33 PM Pavan Rallabhandi 
> <mailto:prallabha...@walmartlabs.com> wrote:
>     Looks like you are running on CentOS, fwiw. We?ve successfully ran the 
> conversion commands on Jewel, Ubuntu 16.04.
>
>     Have a feel it?s expecting the compression to be enabled, can you try 
> removing ?compression=kNoCompression? from the filestore_rocksdb_options? 
> And/or you might want to check if rocksdb is expecting snappy to be enabled.
>
>     From: David Turner <mailto:drakonst...@gmail.com>
>     Date: Tuesday, September 18, 2018 at 6:01 PM
>     To: Pavan Rallabhandi <mailto:prallabha...@walmartlabs.com>
>     Cc: ceph-users <mailto:ceph-users@lists.ceph.com>
>     Subject: EXT: Re: [ceph-users] Any backfill in our cluster makes the 
> cluster unusable and takes forever
>
>     Here's the [1] full log from the time the OSD was started to the end of 
> the crash dump.  These logs are so hard to parse.  Is there anything useful 
> in them?
>
>     I did confirm that all perms were set correctly and that the superblock 
> was changed to rocksdb before the first time I attempted to start the OSD 
> with it's new DB.  This is on a fully Luminous cluster with [2] the defaults 
> you mentioned.
>
>     [1] https://gist.github.com/drakonstein/fa3ac0ad9b2ec1389c957f95e05b79ed
>     [2] "filestore_omap_backend": "rocksdb",
>     "filestore_rocksdb_options": 
> "max_background_compactions=8,compaction_readahead_size=2097152,compression=kNoCompression",
>
>     On Tue, Sep 18, 2018 at 5:29 PM Pavan Rallabhandi 
> <mailto:mailto:prallabha...@walmartlabs.com> wrote:
>     I meant the stack trace hints that the superblock still has leveldb in 
> it, have you verified that already?
>
>     On 9/18/18, 5:27 PM, "Pavan Rallabhandi" 
> <mailto:mailto:prallabha...@walmartlabs.com> wrote:
>
>         You should be able to set them under the global section and that 
> reminds me, since you are on Luminous already, I guess those values are 
> already the default, you can verify from the admin socket of any OSD.
>
>         But the stack trace didn?t hint as if the superblock on the OSD is 
> still considering the omap backend to be leveldb and to do with the 
> compression.
>
>         Thanks,
>         -Pavan.
>
>         From: David Turner <mailto:mailto:drakonst...@gmail.com>
>         Date: Tuesday, September 18, 2018 at 5:07 PM
>         To: Pavan Rallabhandi <mailto:mailto:prallabha...@walmartlabs.com>
>         Cc: ceph-users <mailto:mailto:ceph-users@lists.ceph.com>
>         Subject: EXT: Re: [ceph-users] Any backfill in our cluster makes the 
> cluster unusable and takes forever
>
>         Are those settings fine to have be global even if not all OSDs on a 
> node have rocksdb as the backend?  Or will I need to convert all OSDs on a 
> node at the same time?
>
>         On Tue, Sep 18, 2018 at 5:02 PM Pavan Rallabhandi 
> <mailto:mailto:mailto:mailto:prallabha...@walmartlabs.com> wrote:
>         The steps that were outlined for conversion are correct, have you 
> tried setting some the relevant ceph conf values too:
>
>         filestore_rocksdb_options = 
> "max_background_compactions=8;compaction_readahead_size=2097152;compression=kNoCompression"
>
>         filestore_omap_backend = rocksdb
>
>         Thanks,
>         -Pavan.
>
>         From: ceph-users 
> <mailto:mailto:mailto:mailto:ceph-users-boun...@lists.ceph.com> on behalf of 
> David Turner <mailto:mailto:mailto:mailto:drakonst...@gmail.com>
>         Date: Tuesday, September 18, 2018 at 4:09 PM
>         To: ceph-users <mailto:mailto:mailto:mailto:ceph-users@lists.ceph.com>
>         Subject: EXT: [ceph-users] Any backfill in our cluster makes the 
> cluster unusable and takes forever
>
>         I've finally learned enough about the OSD backend track down this 
> issue to what I believe is the root cause.  LevelDB compaction is the common 
> thread every time we move data around our cluster.  I've ruled out PG 
> subfolder splitting, EC doesn't seem to be the root cause of this, and it is 
> cluster wide as opposed to specific hardware.
>
>         One of the first things I found after digging into leveldb omap 
> compaction was [1] this article with a heading "RocksDB instead of LevelDB" 
> which mentions that leveldb was replaced with rocksdb as the default db 
> backend for filestore OSDs and was even backported to Jewel because of the 
> performance improvements.
>
>         I figured there must be a way to be able to upgrade an OSD to use 
> rocksdb from leveldb without needing to fully backfill the entire OSD.  There 
> is [2] this article, but you need to have an active service account with 
> RedHat to access it.  I eventually came across [3] this article about 
> optimizing Ceph Object Storage which mentions a resolution to OSDs flapping 
> due to omap compaction to migrate to using rocksdb.  It links to the RedHat 
> article, but also has [4] these steps outlined in it.  I tried to follow the 
> steps, but the OSD I tested this on was unable to start with [5] this 
> segfault.  And then trying to move the OSD back to the original LevelDB omap 
> folder resulted in [6] this in the log.  I apologize that all of my logging 
> is with log level 1.  If needed I can get some higher log levels.
>
>         My Ceph version is 12.2.4.  Does anyone have any suggestions for how 
> I can update my filestore backend from leveldb to rocksdb?  Or if that's the 
> wrong direction and I should be looking elsewhere?  Thank you.
>
>
>         [1] https://ceph.com/community/new-luminous-rados-improvements/
>         [2] https://access.redhat.com/solutions/3210951
>         [3] 
> https://hubb.blob.core.windows.net/c2511cea-81c5-4386-8731-cc444ff806df-public/resources/Optimize
>  Ceph object storage for production in multisite clouds.pdf
>
>         [4] ? Stop the OSD
>         ? mv /var/lib/ceph/osd/ceph-/current/omap 
> /var/lib/ceph/osd/ceph-/omap.orig
>         ? ulimit -n 65535
>         ? ceph-kvstore-tool leveldb /var/lib/ceph/osd/ceph-/omap.orig 
> store-copy /var/lib/ceph/osd/ceph-/current/omap 10000 rocksdb
>         ? ceph-osdomap-tool --omap-path /var/lib/ceph/osd/ceph-/current/omap 
> --command check
>         ? sed -i s/leveldb/rocksdb/g /var/lib/ceph/osd/ceph-/superblock
>         ? chown ceph.ceph /var/lib/ceph/osd/ceph-/current/omap -R
>         ? cd /var/lib/ceph/osd/ceph-; rm -rf omap.orig
>         ? Start the OSD
>
>         [5] 2018-09-17 19:23:10.826227 7f1f3f2ab700 -1 abort: Corruption: 
> Snappy not supported or corrupted Snappy compressed block contents
>         2018-09-17 19:23:10.830525 7f1f3f2ab700 -1 *** Caught signal 
> (Aborted) **
>
>         [6] 2018-09-17 19:27:34.010125 7fcdee97cd80 -1 osd.0 0 OSD:init: 
> unable to mount object store
>         2018-09-17 19:27:34.010131 7fcdee97cd80 -1 ESC[0;31m ** ERROR: osd 
> init failed: (1) Operation not permittedESC[0m
>         2018-09-17 19:27:54.225941 7f7f03308d80  0 set uid:gid to 167:167 
> (ceph:ceph)
>         2018-09-17 19:27:54.225975 7f7f03308d80  0 ceph version 12.2.4 
> (52085d5249a80c5f5121a76d6288429f35e4e77b) luminous (stable), process 
> (unknown), pid 361535
>         2018-09-17 19:27:54.231275 7f7f03308d80  0 pidfile_write: ignore 
> empty --pid-file
>         2018-09-17 19:27:54.260207 7f7f03308d80  0 load: jerasure load: lrc 
> load: isa
>         2018-09-17 19:27:54.260520 7f7f03308d80  0 
> filestore(/var/lib/ceph/osd/ceph-0) backend xfs (magic 0x58465342)
>         2018-09-17 19:27:54.261135 7f7f03308d80  0 
> filestore(/var/lib/ceph/osd/ceph-0) backend xfs (magic 0x58465342)
>         2018-09-17 19:27:54.261750 7f7f03308d80  0 
> genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: FIEMAP 
> ioctl is disabled via 'filestore fiemap' config option
>         2018-09-17 19:27:54.261757 7f7f03308d80  0 
> genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: 
> SEEK_DATA/SEEK_HOLE is disabled via 'filestore seek data hole' config option
>         2018-09-17 19:27:54.261758 7f7f03308d80  0 
> genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: splice() 
> is disabled via 'filestore splice' config option
>         2018-09-17 19:27:54.286454 7f7f03308d80  0 
> genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: syncfs(2) 
> syscall fully supported (by glibc and kernel)
>         2018-09-17 19:27:54.286572 7f7f03308d80  0 
> xfsfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_feature: extsize is 
> disabled by conf
>         2018-09-17 19:27:54.287119 7f7f03308d80  0 
> filestore(/var/lib/ceph/osd/ceph-0) start omap initiation
>         2018-09-17 19:27:54.287527 7f7f03308d80 -1 
> filestore(/var/lib/ceph/osd/ceph-0) mount(1723): Error initializing leveldb : 
> Corruption: VersionEdit: unknown tag
>
>
>
>
>
>
> ------------------------------
>
> Subject: Digest Footer
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> ------------------------------
>
> End of ceph-users Digest, Vol 68, Issue 29
> ******************************************
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to