[ceph-users] Cephfs snapshots in Nautilus

2020-05-06 Thread Stolte, Felix
smime.p7m Description: S/MIME encrypted message ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: State of SMR support in Ceph?

2020-05-06 Thread Janne Johansson
Den ons 6 maj 2020 kl 00:58 skrev Oliver Freyermuth < freyerm...@physik.uni-bonn.de>: > Dear Cephalopodians, > seeing the recent moves of major HDD vendors to sell SMR disks targeted > for use in consumer NAS devices (including RAID systems), > I got curious and wonder what the current status of S

[ceph-users] Re: adding block.db to OSD

2020-05-06 Thread Stefan Priebe - Profihost AG
Hi Igor, Am 05.05.20 um 16:10 schrieb Igor Fedotov: > Hi Stefan, > > so (surprise!) some DB access counters show a significant difference, e.g. > > "kv_flush_lat": { > "avgcount": 1423, > "sum": 0.000906419, > "avgtime": 0.00636 > }, >

[ceph-users] Question about bucket versions

2020-05-06 Thread Katarzyna Myrek
Hi I have a few questions about bucket versioning. in the output of command "*radosgw-admin bucket stats --bucket=XXX"* there is info about versions: "ver": "0#521391,1#516042,2#518098,3#517681,4#518423", "master_ver": "0#0,1#0,2#0,3#0,4#0", Also "*metadata get"* returns info about vers

[ceph-users] Re: Cephfs snapshots in Nautilus

2020-05-06 Thread Marc Roos
I have been using snapshots on cephfs since luminous, 1xfs and 1xactivemds and used an rsync on it for backup. Under luminious I did not encounter any problems with this setup. I think I was even snapshotting user dirs every 7 days having thousands of snapshots (which I later heard, is not rec

[ceph-users] nautilus cluster not dynamically resharding

2020-05-06 Thread Marcel Ceph
Hi all, I have 3 nautilus clusters. They started out as mimic but were recently upgraded to nautilus. On two of them dynamic bucket index sharding seems to work automatically. However on one of the clusters it doesn;t and I have no clue why * If I execute radosgw-admin bucket limit check I find s

[ceph-users] Cephfs snapshots Nautilus

2020-05-06 Thread Stolte, Felix
Hi Folks, I really like to use snapshots on cephfs, but even on octopus release snapshots are still marked as an experimental feature. Is anyone using snapshots in production environments? Which issues did you encounter? Do I risk a corrupted filesystem or just non-working snapshots? We run

[ceph-users] Fwd: Octopus on CentOS 7: lacking some packages

2020-05-06 Thread Sam Huracan
Hi Cephers, I am trying to install Ceph Octopus using ceph-deploy on CentOS 7, as installing ceph-mgr-dashboard, it required these packages: - python3-cherrypy - python3-jwt - python3-routes https://pastebin.com/dSQPgGJD But, when I installed these packages, they were not available. https://pa

[ceph-users] Re: Cephfs snapshots in Nautilus

2020-05-06 Thread Yan, Zheng
On Wed, May 6, 2020 at 3:53 PM Marc Roos wrote: > > > I have been using snapshots on cephfs since luminous, 1xfs and > 1xactivemds and used an rsync on it for backup. > Under luminious I did not encounter any problems with this setup. I > think I was even snapshotting user dirs every 7 days having

[ceph-users] Re: Fwd: Octopus on CentOS 7: lacking some packages

2020-05-06 Thread Ashley Merrick
As per the release notes : https://docs.ceph.com/docs/master/releases/octopus/ The dashboard and a few other modules aren't supported on CentOS 7.x due to python version / dependencies. On Wed, 06 May 2020 17:18:06 +0800 Sam Huracan wrote Hi Cephers, I am trying to install C

[ceph-users] Re: Cephfs snapshots in Nautilus

2020-05-06 Thread Stolte, Felix
Hi Marc, thank you for the response. Since upgrading to nautilus I also encounter the cache pressure warnings while doing tape backup of the filesystem, so it is probably not related to snapshots. Anyone else who likes to share his experience with cephfs snapshots? Best regards Felix

[ceph-users] Re: State of SMR support in Ceph?

2020-05-06 Thread Oliver Freyermuth
Dear Janne, Am 06.05.20 um 09:18 schrieb Janne Johansson: Den ons 6 maj 2020 kl 00:58 skrev Oliver Freyermuth mailto:freyerm...@physik.uni-bonn.de>>: Dear Cephalopodians, seeing the recent moves of major HDD vendors to sell SMR disks targeted for use in consumer NAS devices (including

[ceph-users] Re: radosgw garbage collection error

2020-05-06 Thread Pritha Srivastava
From these: crt=5395'481460 lcod 5395'481461 mlcod 5395'481461 active+clean] do_osd_op sync_read 4096~1024 2020-05-06T08:06:33.925+0200 7f73b554a700 10 osd.15 pg_epoch: 5395 pg[5.9( v 5395'481462 (5387'478000,5395'481462] local-lis/les=5394/5395 n=48 ec=67/67 lis/c=5394/5394 les/c/f=5395/5395/0 sis

[ceph-users] Re: State of SMR support in Ceph?

2020-05-06 Thread Cranage, Steve
We’re prototyping a native SMR object store now to run alongside Ceph, which has been our only object store backend for the last 3 years and is somewhat problematic on some metrics. I believe trying to use SMR drives with a file system in the architecture (as in Ceph) is a non starter, the solut

[ceph-users] Re: State of SMR support in Ceph?

2020-05-06 Thread brad . swanson
Take a look at the available SMR drives: https://www.ixsystems.com/community/resources/list-of-known-smr-drives.141/ I wouldn’t put a single one of those drives into a Ceph system. You won’t save any money. In fact, it’s likely a sink-hole for labor on your part. In fact, you couldn’t pay me

[ceph-users] Re: State of SMR support in Ceph?

2020-05-06 Thread Oliver Freyermuth
Am 06.05.20 um 16:12 schrieb brad.swan...@adtran.com: Take a look at the available SMR drives: https://www.ixsystems.com/community/resources/list-of-known-smr-drives.141/ Thanks for this nice overview link! Indeed, my question was more thinking about "the next years". While currently, SMR is c

[ceph-users] RBD throughput/IOPS benchmarks

2020-05-06 Thread Vincent KHERBACHE
Dear all, We currently run a small Ceph cluster on 2 machines and we wonder what are the theoretical max BW/IOPS we can achieve through RBD with our setup. Here are the environment details: - The Ceph release is an octopus 15.2.1 running on Centos 8, both machines have 180GB RAM, 72 cores, a

[ceph-users] Re: Ceph meltdown, need help

2020-05-06 Thread Marc Roos
Made it all the way down ;) Thank you very much for the detailed info. -Original Message- Cc: ceph-users Subject: Re: [ceph-users] Re: Ceph meltdown, need help Dear Marc, This e-mail is two-part. First part is about the problem of a single client being able to crash a ceph cluster. Y

[ceph-users] Re: How many MDS servers

2020-05-06 Thread Patrick Donnelly
Hello Robert, On Mon, Mar 9, 2020 at 7:55 PM Robert Ruge wrote: > For a 1.1PB raw cephfs system currently storing 191TB of data and 390 million > objects (mostly small Python, ML training files etc.) how many MDS servers > should I be running? > > System is Nautilus 14.2.8. > > > > I ask becaus

[ceph-users] Re: ceph: Can't lookup inode 1 (err: -13)

2020-05-06 Thread Patrick Donnelly
On Mon, Mar 9, 2020 at 3:19 PM Marc Roos wrote: > > > For testing purposes I changed the kernel 3.10 for a 5.5, now I am > getting these messages. I assume the 3.10 was just never displaying > these. Could this be a problem with my caps of the fs id user? > > [Mon Mar 9 23:10:52 2020] ceph: Can't

[ceph-users] Re: Data loss by adding 2OSD causing Long heartbeat ping times

2020-05-06 Thread Frank Schilder
To answer some of my own questions: 1) Setting ceph osd set noout ceph osd set nodown ceph osd set norebalance before restart/re-deployment did not harm. I don't know if it helped, because I didn't retry the procedure that led to OSDs going down. See also point 3 below. 2) A peculiarity of thi

[ceph-users] Error with zabbix module on Ceph Octopus

2020-05-06 Thread Matt Larson
Hi, I am trying to setup the Zabbix reporting module, but it is giving an error which looks like a Python error: ceph zabbix config-show Error EINVAL: TypeError: __init__() got an unexpected keyword argument 'index' I have configured the zabbix_host and identifier already at this point. The co

[ceph-users] Re: radosgw garbage collection error

2020-05-06 Thread James, GleSYS
Hi, Initial installation was 15.2.0 and we’ve upgraded to 15.2.1. I’ve verified and the RGWs and OSDs are running the same version - 15.2.1. I’ve created a bug report for this issue and uploaded the debug logs as requested: https://tracker.ceph.com/issues/45412

[ceph-users] Re: CephFS with active-active NFS Ganesha

2020-05-06 Thread Patrick Donnelly
Hello Michael, On Wed, Mar 11, 2020 at 1:24 AM Michael Bisig wrote: > > Hi all, > > I am trying to setup an active-active NFS Ganesha cluster (with two Ganeshas > (v3.0) running in Docker containers). I could manage to get two Ganesha > daemons running using the rados_cluster backend for active

[ceph-users] Re: Ceph meltdown, need help

2020-05-06 Thread Frank Schilder
Dear Marc, This e-mail is two-part. First part is about the problem of a single client being able to crash a ceph cluster. Yes, I think this applies to many, if not any cluster. Second part is about your question about the time-out value. Part 1: Yes, I think there is a serious problem with ce

[ceph-users] Re: Cluster blacklists MDS, can't start

2020-05-06 Thread Patrick Donnelly
On Wed, Mar 11, 2020 at 10:41 PM Robert LeBlanc wrote: > > This is the second time this happened in a couple of weeks. The MDS locks > up and the stand-by can't take over so the Montiors black list them. I try > to unblack list them, but they still say this in the logs > > mds.0.1184394 waiting fo

[ceph-users] Re: Workload in Unit testing

2020-05-06 Thread Bobby
Hi Sebastian, Thanks a lot for your reply. It was really helpful and it is clear that 'make check' don't start a ceph cluster. After your email I have figured it out. This brings me to another question :-) In my earlier email I should have defined what exactly I mean by 'workload' in my case. Giv

[ceph-users] Re: CephFS with active-active NFS Ganesha

2020-05-06 Thread Jeff Layton
On Wed, 2020-05-06 at 11:58 -0700, Patrick Donnelly wrote: > Hello Michael, > > On Wed, Mar 11, 2020 at 1:24 AM Michael Bisig wrote: > > Hi all, > > > > I am trying to setup an active-active NFS Ganesha cluster (with two > > Ganeshas (v3.0) running in Docker containers). I could manage to get t

[ceph-users] Re: What's the best practice for Erasure Coding

2020-05-06 Thread Frank Schilder
Dear Alex, I don't really have a reference for this set up. The ceph documentation describes this as the simplest possible set up and back then it was basically dictated by budget. Everything else was several months of experimentation and benchmarking. I had scripts running for several weeks ju

[ceph-users] per rbd performance counters

2020-05-06 Thread Void Star Nill
Hello, Is there a way to get read/write I/O statistics for each rbd device for each mapping? For example, when an application uses one of the volumes, I would like to find out what performance (avg read/write bandwidth, IOPS, etc) that application observed on a given volume. Is that possible? Th

[ceph-users] Re: per rbd performance counters

2020-05-06 Thread Peter Parker
Hi Shridhar, Did u mean 'ceph rbd perf image stats' or  'rbd perf image iostat' ? for monitor intergration,  you can enable mgr prometheus module and get metrics for each image in specified pools -- Original -- From: "Void Star Nill"

[ceph-users] changed caps not propagated to kernel cephfs mounts

2020-05-06 Thread Andrej Filipcic
Hi, I have added a new fast_data pool to cephfs, fixed the auth caps, eg client.f9wn     key:     caps: [mds] allow rw     caps: [mon] allow r     caps: [osd] allow rw pool=cephfs_data, allow rw pool=fast_data but the client with kernel mounted cephfs reports an error w

[ceph-users] Re: rados buckets copy

2020-05-06 Thread Szabo, Istvan (Agoda)
Hi, You might try s3 browser app, it is quite easy to navigate and copy between buckets. Istvan Szabo Senior Infrastructure Engineer --- Agoda Services Co., Ltd. e: istvan.sz...@agoda.com --- -Or