[ceph-users] Using multisite to migrate data between bucket data pools.

2019-10-30 Thread David Turner
This is a tangent on Paul Emmerich's response to "[ceph-users] Correct Migration Workflow Replicated -> Erasure Code". I've tried Paul's method before to migrate between 2 data pools. However I ran into some issues. The first issue seems like a bug in RGW where the RGW for the new zone was able to

Re: [ceph-users] ceph balancer do not start

2019-10-22 Thread David Turner
Of the top of my head, if say your cluster might have the wrong tunables for crush-compat. I know I ran into that when I first set up the balancer and nothing obviously said that was the problem. Only researching find it for me. My real question, though, is why aren't you using upmap? It is signif

Re: [ceph-users] Decreasing the impact of reweighting osds

2019-10-22 Thread David Turner
Most times you are better served with simpler settings like osd_recovery_sleep, which has 3 variants if you have multiple types of OSDs in your cluster (osd_recovery_sleep_hdd, osd_recovery_sleep_sdd, osd_recovery_sleep_hybrid). Using those you can tweak a specific type of OSD that might be having

Re: [ceph-users] Cannot delete bucket

2019-06-27 Thread David Turner
lly help with bucket deletion. On Thu, Jun 27, 2019 at 2:58 PM Sergei Genchev wrote: > @David Turner > Did your bucket delete ever finish? I am up to 35M incomplete uploads, > and I doubt that I actually had that many upload attempts. I could be > wrong though. > Is there a way to

Re: [ceph-users] Cannot delete bucket

2019-06-24 Thread David Turner
It's aborting incomplete multipart uploads that were left around. First it will clean up the cruft like that and then it should start actually deleting the objects visible in stats. That's my understanding of it anyway. I'm int he middle of cleaning up some buckets right now doing this same thing.

Re: [ceph-users] Changing the release cadence

2019-06-17 Thread David Turner
This was a little long to respond with on Twitter, so I thought I'd share my thoughts here. I love the idea of a 12 month cadence. I like October because admins aren't upgrading production within the first few months of a new release. It gives it plenty of time to be stable for the OS distros as we

Re: [ceph-users] Default Pools

2019-04-23 Thread David Turner
You should be able to see all pools in use in a RGW zone from the radosgw-admin command. This [1] is probably overkill for most, but I deal with multi-realm clusters so I generally think like this when dealing with RGW. Running this as is will create a file in your current directory for each zone

Re: [ceph-users] Osd update from 12.2.11 to 12.2.12

2019-04-22 Thread David Turner
Do you perhaps have anything in the ceph.conf files on the servers with those OSDs that would attempt to tell the daemon that they are filestore osds instead of bluestore? I'm sure you know that the second part [1] of the output in both cases only shows up after an OSD has been rebooted. I'm sure

Re: [ceph-users] ceph osd pg-upmap-items not working

2019-03-15 Thread David Turner
Why do you think that it can't resolve this by itself? You just said that the balancer was able to provide an optimization, but then that the distribution isn't perfect. When there are no further optimizations, running `ceph balancer optimize plan` won't create a plan with any changes. Possibly

Re: [ceph-users] OpenStack with Ceph RDMA

2019-03-11 Thread David Turner
I can't speak to the rdma portion. But to clear up what each of these does... the cluster network is only traffic between the osds for replicating writes, reading EC data, as well as backfilling and recovery io. Mons, mds, rgw, and osds talking with clients all happen on the public network. The gen

Re: [ceph-users] priorize degraged objects than misplaced

2019-03-11 Thread David Turner
Ceph has been getting better and better about prioritizing this sorry of recovery, but free of those optimizations are in Jewel, which had been out of the support cycle for about a year. You should look into upgrading to mimic where you should see a pretty good improvement on this sorry of prioriti

Re: [ceph-users] CEPH ISCSI Gateway

2019-03-11 Thread David Turner
The problem with clients on osd nodes is for kernel clients only. That's true of krbd and the kernel client for cephfs. The only other reason not to run any other Ceph daemon in the same node as osds is resource contention if you're running at higher CPU and memory utilizations. On Sat, Mar 9, 201

Re: [ceph-users] rbd unmap fails with error: rbd: sysfs write failed rbd: unmap failed: (16) Device or resource busy

2019-03-01 Thread David Turner
n Fri, Mar 1, 2019, 6:28 PM solarflow99 wrote: > It has to be mounted from somewhere, if that server goes offline, you need > to mount it from somewhere else right? > > > On Thu, Feb 28, 2019 at 11:15 PM David Turner > wrote: > >> Why are you making the same rbd to multip

Re: [ceph-users] Mimic 13.2.4 rbd du slowness

2019-02-28 Thread David Turner
Have you used strace on the du command to see what it's spending its time doing? On Thu, Feb 28, 2019, 8:45 PM Glen Baars wrote: > Hello Wido, > > The cluster layout is as follows: > > 3 x Monitor hosts ( 2 x 10Gbit bonded ) > 9 x OSD hosts ( > 2 x 10Gbit bonded, > LSI cachecade and write cache

Re: [ceph-users] rbd unmap fails with error: rbd: sysfs write failed rbd: unmap failed: (16) Device or resource busy

2019-02-28 Thread David Turner
Why are you making the same rbd to multiple servers? On Wed, Feb 27, 2019, 9:50 AM Ilya Dryomov wrote: > On Wed, Feb 27, 2019 at 12:00 PM Thomas <74cmo...@gmail.com> wrote: > > > > Hi, > > I have noticed an error when writing to a mapped RBD. > > Therefore I unmounted the block device. > > Then

Re: [ceph-users] PG Calculations Issue

2019-02-28 Thread David Turner
Those numbers look right for a pool only containing 10% of your data. Now continue to calculate the pg counts for the remaining 90% of your data. On Wed, Feb 27, 2019, 12:17 PM Krishna Venkata wrote: > Greetings, > > > I am having issues in the way PGs are calculated in > https://ceph.com/pgcalc

Re: [ceph-users] redirect log to syslog and disable log to stderr

2019-02-28 Thread David Turner
You can always set it in your ceph.conf file and restart the mgr daemon. On Tue, Feb 26, 2019, 1:30 PM Alex Litvak wrote: > Dear Cephers, > > In mimic 13.2.2 > ceph tell mgr.* injectargs --log-to-stderr=false > Returns an error (no valid command found ...). What is the correct way to > inject m

Re: [ceph-users] Right way to delete OSD from cluster?

2019-02-28 Thread David Turner
The reason is that an osd still contributes to the host weight in the crush map even while it is marked out. When you out and then purge, the purging operation removed the osd from the map and changes the weight of the host which changes the crush map and data moves. By weighting the osd to 0.0, th

[ceph-users] Usenix Vault 2019

2019-02-24 Thread David Turner
There is a scheduled birds of a feather for Ceph tomorrow night, but I also noticed that there are only trainings tomorrow. Unless you are paying more for those, you likely don't have much to do on Monday. That's the boat I'm in. Is anyone interested in getting together tomorrow in Boston during th

Re: [ceph-users] Configuration about using nvme SSD

2019-02-24 Thread David Turner
One thing that's worked for me to get more out of nvmes with Ceph is to create multiple partitions on the nvme with an osd on each partition. That way you get more osd processes and CPU per nvme device. I've heard of people using up to 4 partitions like this. On Sun, Feb 24, 2019, 10:25 AM Vitaliy

Re: [ceph-users] Doubts about backfilling performance

2019-02-23 Thread David Turner
Jewel is really limited on the settings you can tweak for backfilling [1]. Luminous and Mimic have a few more knobs. An option you can do, though, is to use osd_crush_initial_weight found [2] here. With this setting you set your initial crush weight for new osds to 0.0 and gradually increase them t

Re: [ceph-users] Ceph cluster stability

2019-02-22 Thread David Turner
luestore that does > >> > it, but I'm seeing lower cpu usage when moving to bluestore + rocksdb > >> > compared to filestore + leveldb . > >> > > >> > > >> > On Wed, Feb 20, 2019 at 4:27 PM M Ranga Swami Reddy > >> > wrot

Re: [ceph-users] REQUEST_SLOW across many OSDs at the same time

2019-02-22 Thread David Turner
Can you correlate the times to scheduled tasks inside of any VMs? For instance if you have several Linux VMs with the updatedb command installed that by default they will all be scanning their disks at the same time each day to see where files are. Other common culprits could be scheduled backups,

Re: [ceph-users] Ceph cluster stability

2019-02-22 Thread David Turner
when moving to bluestore + rocksdb > > compared to filestore + leveldb . > > > > > > On Wed, Feb 20, 2019 at 4:27 PM M Ranga Swami Reddy > > wrote: > > > > > > Thats expected from Ceph by design. But in our case, we are using all > > > recomme

Re: [ceph-users] faster switch to another mds

2019-02-20 Thread David Turner
If I'm not mistaken, if you stop them at the same time during a reboot on a node with both mds and mon, the mons might receive it, but wait to finish their own election vote before doing anything about it. If you're trying to keep optimal uptime for your mds, then stopping it first and on its own

Re: [ceph-users] faster switch to another mds

2019-02-19 Thread David Turner
It's also been mentioned a few times that when MDS and MON are on the same host that the downtime for MDS is longer when both daemons stop at about the same time. It's been suggested to stop the MDS daemon, wait for `ceph mds stat` to reflect the change, and then restart the rest of the server. HT

Re: [ceph-users] CephFS overwrite/truncate performance hit

2019-02-19 Thread David Turner
If your client needs to be able to handle the writes like that on its own, RBDs might be the more appropriate use case. You lose the ability to have multiple clients accessing the data as easily as with CephFS, but you would gain the features you're looking for. On Tue, Feb 12, 2019 at 1:43 PM Gr

Re: [ceph-users] CephFS: client hangs

2019-02-19 Thread David Turner
You're attempting to use mismatching client name and keyring. You want to use matching name and keyring. For your example, you would want to either use `--keyring /etc/ceph/ceph.client.admin.keyring --name client.admin` or `--keyring /etc/ceph/ceph.client.cephfs.keyring --name client.cephfs`. Mix

Re: [ceph-users] crush map has straw_calc_version=0 and legacy tunables on luminous

2019-02-19 Thread David Turner
[1] Here is a really cool set of slides from Ceph Day Berlin where Dan van der Ster uses the mgr balancer module with upmap to gradually change the tunables of a cluster without causing major client impact. The down side for you is that upmap requires all luminous or newer clients, but if you upgr

Re: [ceph-users] Ceph cluster stability

2019-02-19 Thread David Turner
With a RACK failure domain, you should be able to have an entire rack powered down without noticing any major impact on the clients. I regularly take down OSDs and nodes for maintenance and upgrades without seeing any problems with client IO. On Tue, Feb 12, 2019 at 5:01 AM M Ranga Swami Reddy w

Re: [ceph-users] Migrating a baremetal Ceph cluster into K8s + Rook

2019-02-19 Thread David Turner
et this up and running aswell. (But keep in > mind that the help on their mailing list is not so good as here ;)) > > > > -Original Message- > From: David Turner [mailto:drakonst...@gmail.com] > Sent: 18 February 2019 17:31 > To: ceph-users > Subject: [ceph-users]

Re: [ceph-users] [Bluestore] Some of my osd's uses BlueFS slow storage for db - why?

2019-02-19 Thread David Turner
yone to migrate their DB back off of their spinner which is what's happening to you. I don't believe that sort of tooling exists yet, though, without compiling the Nautilus Beta tooling for yourself. On Tue, Feb 19, 2019 at 12:03 AM Konstantin Shalygin wrote: > On 2/18/19 9:43 PM, Da

Re: [ceph-users] Upgrade Luminous to mimic on Ubuntu 18.04

2019-02-18 Thread David Turner
Everybody is just confused that you don't have a newer version of Ceph available. Are you running `apt-get dist-upgrade` to upgrade ceph? Do you have any packages being held back? There is no reason that Ubuntu 18.04 shouldn't be able to upgrade to 12.2.11. On Mon, Feb 18, 2019, 4:38 PM Hello peo

Re: [ceph-users] IRC channels now require registered and identified users

2019-02-18 Thread David Turner
Is this still broken in the 1-way direction where Slack users' comments do not show up in IRC? That would explain why nothing I ever type (as either helping someone or asking a question) ever have anyone respond to them. On Tue, Dec 18, 2018 at 6:50 AM Joao Eduardo Luis wrote: > On 12/18/2018 1

[ceph-users] Migrating a baremetal Ceph cluster into K8s + Rook

2019-02-18 Thread David Turner
I'm getting some "new" (to me) hardware that I'm going to upgrade my home Ceph cluster with. Currently it's running a Proxmox cluster (Debian) which precludes me from upgrading to Mimic. I am thinking about taking the opportunity to convert most of my VMs into containers and migrate my cluster in

[ceph-users] Intel P4600 3.2TB U.2 form factor NVMe firmware problems causing dead disks

2019-02-18 Thread David Turner
We have 2 clusters of [1] these disks that have 2 Bluestore OSDs per disk (partitioned), 3 disks per node, 5 nodes per cluster. The clusters are 12.2.4 running CephFS and RBDs. So in total we have 15 NVMe's per cluster and 30 NVMe's in total. They were all built at the same time and were running

Re: [ceph-users] Placing replaced disks to correct buckets.

2019-02-18 Thread David Turner
Also what commands did you run to remove the failed HDDs and the commands you have so far run to add their replacements back in? On Sat, Feb 16, 2019 at 9:55 PM Konstantin Shalygin wrote: > I recently replaced failed HDDs and removed them from their respective > buckets as per procedure. > > But

Re: [ceph-users] [Bluestore] Some of my osd's uses BlueFS slow storage for db - why?

2019-02-18 Thread David Turner
been there ever since. On Sat, Feb 16, 2019 at 1:50 AM Konstantin Shalygin wrote: > On 2/16/19 12:33 AM, David Turner wrote: > > The answer is probably going to be in how big your DB partition is vs > > how big your HDD disk is. From your output it looks like you have a > >

Re: [ceph-users] [Bluestore] Some of my osd's uses BlueFS slow storage for db - why?

2019-02-15 Thread David Turner
The answer is probably going to be in how big your DB partition is vs how big your HDD disk is. From your output it looks like you have a 6TB HDD with a 28GB Blocks.DB partition. Even though the DB used size isn't currently full, I would guess that at some point since this OSD was created that it

Re: [ceph-users] jewel10.2.11 EC pool out a osd, its PGs remap to the osds in the same host

2019-02-15 Thread David Turner
I'm leaving the response on the CRUSH rule for Gregory, but you have another problem you're running into that is causing more of this data to stay on this node than you intend. While you `out` the OSD it is still contributing to the Host's weight. So the host is still set to receive that amount o

Re: [ceph-users] Problems with osd creation in Ubuntu 18.04, ceph 13.2.4-1bionic

2019-02-15 Thread David Turner
I have found that running a zap before all prepare/create commands with ceph-volume helps things run smoother. Zap is specifically there to clear everything on a disk away to make the disk ready to be used as an OSD. Your wipefs command is still fine, but then I would lvm zap the disk before conti

Re: [ceph-users] [Ceph-community] Deploy and destroy monitors

2019-02-13 Thread David Turner
Ceph-users is the proper ML to post questions like this. On Thu, Dec 20, 2018 at 2:30 PM Joao Eduardo Luis wrote: > On 12/20/2018 04:55 PM, João Aguiar wrote: > > I am having an issue with "ceph-ceploy mon” > > > > I started by creating a cluster with one monitor with "create-deploy > new"… "cre

Re: [ceph-users] [Ceph-community] Ceph SSE-KMS integration to use Safenet as Key Manager service

2019-02-13 Thread David Turner
Ceph-users is the correct ML to post questions like this. On Wed, Jan 2, 2019 at 5:40 PM Rishabh S wrote: > Dear Members, > > Please let me know if you have any link with examples/detailed steps of > Ceph-Safenet(KMS) integration. > > Thanks & Regards, > Rishabh > > _

Re: [ceph-users] [Ceph-community] Error during playbook deployment: TASK [ceph-mon : test if rbd exists]

2019-02-13 Thread David Turner
Ceph-users ML is the proper mailing list for questions like this. On Sat, Jan 26, 2019 at 12:31 PM Meysam Kamali wrote: > Hi Ceph Community, > > I am using ansible 2.2 and ceph branch stable-2.2, on centos7, to deploy > the playbook. But the deployment get hangs in this step "TASK [ceph-mon : >

Re: [ceph-users] [Ceph-community] Need help related to ceph client authentication

2019-02-13 Thread David Turner
The Ceph-users ML is the correct list to ask questions like this. Did you figure out the problems/questions you had? On Tue, Dec 4, 2018 at 11:39 PM Rishabh S wrote: > Hi Gaurav, > > Thank You. > > Yes, I am using boto, though I was looking for suggestions on how my ceph > client should get acc

Re: [ceph-users] all vms can not start up when boot all the ceph hosts.

2019-02-13 Thread David Turner
This might not be a Ceph issue at all depending on if you're using any sort of caching. If you have caching on your disk controllers at all, then the write might have happened to the cache but never made it to the OSD disks which would show up as problems on the VM RBDs. Make sure you have proper

Re: [ceph-users] how to mount one of the cephfs namespace using ceph-fuse?

2019-02-13 Thread David Turner
Note that this format in fstab does require a certain version of util-linux because of the funky format of the line. Pretty much it maps all command line options at the beginning of the line separated with commas. On Wed, Feb 13, 2019 at 2:10 PM David Turner wrote: > I believe the fstab l

Re: [ceph-users] how to mount one of the cephfs namespace using ceph-fuse?

2019-02-13 Thread David Turner
I believe the fstab line for ceph-fuse in this case would look something like [1] this. We use a line very similar to that to mount cephfs at a specific client_mountpoint that the specific cephx user only has access to. [1] id=acapp3,client_mds_namespace=fs1 /tmp/ceph fuse.ceph defaults,noat

Re: [ceph-users] compacting omap doubles its size

2019-02-13 Thread David Turner
Sorry for the late response on this, but life has been really busy over the holidays. We compact our omaps offline with the ceph-kvstore-tool. Here [1] is a copy of the script that we use for our clusters. You might need to modify things a bit for your environment. I don't remember which versio

Re: [ceph-users] can not start osd service by systemd

2018-11-19 Thread David Turner
I believe I fixed this issue by running `systemctl enable ceph-osd@n.service` for all of the OSDs and then it wasn't a problem in future. On Fri, Nov 9, 2018 at 9:30 PM wrote: > Hi! > > I find a confused question about start/stop ceph cluster by systemd: > > - when cluster is on, restart ceph.ta

Re: [ceph-users] Migrate OSD journal to SSD partition

2018-11-19 Thread David Turner
For this the procedure is generally to stop the osd, flush the journal, update the symlink on the osd to the new journal location, mkjournal, start osd. You shouldn't need to do anything in the ceph.conf file. On Thu, Nov 8, 2018 at 2:41 AM wrote: > Hi all, > > > > I have been trying to migrate

Re: [ceph-users] Mimic - EC and crush rules - clarification

2018-11-16 Thread David Turner
while only using as much space as 2x replication. On Thu, Nov 1, 2018 at 11:25 PM Wladimir Mutel wrote: > David Turner wrote: > > Yes, when creating an EC profile, it automatically creates a CRUSH rule > > specific for that EC profile. You are also correct that 2+1 doesn&#

Re: [ceph-users] ceph 12.2.9 release

2018-11-07 Thread David Turner
My big question is that we've had a few of these releases this year that are bugged and shouldn't be upgraded to... They don't have any release notes or announcement and the only time this comes out is when users finally ask about it weeks later. Why is this not proactively announced to avoid a pr

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-05 Thread David Turner
ic version of the tool to update my omap backends. On Mon, Nov 5, 2018 at 4:26 PM Pavan Rallabhandi < prallabha...@walmartlabs.com> wrote: > Not sure I understand that, but starting Luminous, the filestore omap > backend is rocksdb by default. > > > > *From: *David Turn

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-05 Thread David Turner
y, this feature was supported in Jewel starting 10.2.11, ref > https://github.com/ceph/ceph/pull/18010 > > > > I thought you mentioned you were using Luminous 12.2.4. > > > > *From: *David Turner > *Date: *Friday, November 2, 2018 at 5:21 PM > > > *To: *Pavan Ra

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-02 Thread David Turner
other tools as well that can be used to set up the omap backend db. On Fri, Nov 2, 2018, 4:26 PM Pavan Rallabhandi wrote: > It was Redhat versioned Jewel. But may be more relevantly, we are on > Ubuntu unlike your case. > > > > *From: *David Turner > *Date: *Friday, No

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-11-02 Thread David Turner
. > > Thanks, > -Pavan. > > From: David Turner > Date: Monday, October 1, 2018 at 1:37 PM > To: Pavan Rallabhandi > Cc: ceph-users > Subject: EXT: Re: [ceph-users] Any backfill in our cluster makes the > cluster unusable and takes forever > > I tried m

Re: [ceph-users] Mimic - EC and crush rules - clarification

2018-11-01 Thread David Turner
Yes, when creating an EC profile, it automatically creates a CRUSH rule specific for that EC profile. You are also correct that 2+1 doesn't really have any resiliency built in. 2+2 would allow 1 node to go down while still having your data accessible. It will use 2x data to raw as opposed to the

Re: [ceph-users] Packages for debian in Ceph repo

2018-10-30 Thread David Turner
What version of qemu-img are you using? I found [1] this when poking around on my qemu server when checking for rbd support. This version (note it's proxmox) has rbd listed as a supported format. [1] # qemu-img -V; qemu-img --help|grep rbd qemu-img version 2.11.2pve-qemu-kvm_2.11.2-1 Copyright (

Re: [ceph-users] Balancer module not balancing perfectly

2018-10-30 Thread David Turner
>From the balancer module's code for v 12.2.7 I noticed [1] these lines which reference [2] these 2 config options for upmap. You might try using more max iterations or a smaller max deviation to see if you can get a better balance in your cluster. I would try to start with [3] these commands/value

Re: [ceph-users] OSD node reinstallation

2018-10-30 Thread David Turner
otstrap-osd/ceph.keyring >> >> --- >> Alex >> >> On Tue, Oct 30, 2018 at 4:48 AM David Turner >> wrote: >> > >> > Set noout, reinstall the OS without going the OSDs (including any >> journal partitions and maintaining any dmcrypt keys if yo

Re: [ceph-users] OSD node reinstallation

2018-10-29 Thread David Turner
Set noout, reinstall the OS without going the OSDs (including any journal partitions and maintaining any dmcrypt keys if you have encryption), install ceph, make sure the ceph.conf file is correct,zip start OSDs, unset noout once they're back up and in. All of the data the OSD needs to start is on

Re: [ceph-users] reducing min_size on erasure coded pool may allow recovery ?

2018-10-29 Thread David Turner
min_size should be at least k+1 for EC. There are times to use k for emergencies like you had. I would suggest seeing it back to 3 once your back to healthy. As far as why you needed to reduce min_size, my guess would be that recovery would have happened as long as k copies were up. Were the PG's

Re: [ceph-users] Need advise on proper cluster reweighing

2018-10-28 Thread David Turner
Which version of Ceph are you running? Do you have any kernel clients? If yes, can which version kernel? These questions are all leading to see if you can enable the Luminous/Mimic mgr module balancer with upmap. If you can, it is hands down the best way to balance your cluster. On Sat, Oct 27, 20

Re: [ceph-users] Verifying the location of the wal

2018-10-28 Thread David Turner
If your had a specific location for the wal it would show up there. If there is no entry for the wal, then it is using the same seeing as the db. On Sun, Oct 28, 2018, 9:26 PM Robert Stanford wrote: > > Mehmet: it doesn't look like wal is mentioned in the osd metadata. I see > bluefs slow, blu

Re: [ceph-users] Migrate/convert replicated pool to EC?

2018-10-26 Thread David Turner
It is indeed adding a placement target and not removing it replacing the pool. The get/put wouldn't be a rados or even ceph command, you would do it through an s3 client. On Fri, Oct 26, 2018, 9:38 AM Matthew Vernon wrote: > Hi, > > On 26/10/2018 12:38, Alexandru Cucu wrote: > > > Have a look at

Re: [ceph-users] RGW: move bucket from one placement to another

2018-10-25 Thread David Turner
Resharding a bucket won't affect the data in the bucket. After you change the placement for a bucket, you could update where the data is by re-writing all of the data in the bucket. On Thu, Oct 25, 2018 at 8:48 AM Jacek Suchenia wrote: > Hi > > We have a bucket created with LocationConstraint s

Re: [ceph-users] Migrate/convert replicated pool to EC?

2018-10-25 Thread David Turner
There are no tools to migrate in either direction between EC and Replica. You can't even migrate an EC pool to a new EC profile. With RGW you can create a new data pool and new objects will be written to the new pool. If your objects have a lifecycle, then eventually you'll be to the new pool over

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
he DB + WAL on the NVMe? Is this "data disk" > still an ssd? > > > > On Mon, Oct 22, 2018 at 3:34 PM David Turner > wrote: > >> And by the data disk I mean that I didn't specify a location for the DB >> partition. >> >> On Mon, Oct 22

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
ere is more utilization > on the cluster than what's in the pools, the excess equal to about wal size > * number of osds... > > On Mon, Oct 22, 2018 at 3:35 PM David Turner > wrote: > >> My DB doesn't have a specific partition anywhere, but there's still a >&g

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
master/rados/configuration/bluestore-config-ref/ > > On Mon, Oct 22, 2018 at 3:13 PM Robert Stanford > wrote: > >> >> We're out of sync, I think. You have your DB on your data disk so your >> block.db symlink points to that disk, right? There is however no wal

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
db points? I assumed that was a symlink for the db, but necessarily > for the wal, because it can live in a place different than the db. > > On Mon, Oct 22, 2018 at 2:18 PM David Turner > wrote: > >> You can always just go to /var/lib/ceph/osd/ceph-{osd-num}/ and look at >&

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
And by the data disk I mean that I didn't specify a location for the DB partition. On Mon, Oct 22, 2018 at 4:06 PM David Turner wrote: > Track down where it says they point to? Does it match what you expect? > It does for me. I have my DB on my data disk and my WAL on a separate N

Re: [ceph-users] Drive for Wal and Db

2018-10-22 Thread David Turner
You can always just go to /var/lib/ceph/osd/ceph-{osd-num}/ and look at where the symlinks for block and block.wal point to. On Mon, Oct 22, 2018 at 12:29 PM Robert Stanford wrote: > > That's what they say, however I did exactly this and my cluster > utilization is higher than the total pool ut

Re: [ceph-users] ceph df space usage confusion - balancing needed?

2018-10-22 Thread David Turner
I haven't had crush-compat do anything helpful for balancing my clusters. upmap has been amazing and balanced my clusters far better than anything else I've ever seen. I would go so far as to say that upmap can achieve a perfect balance. It seems to evenly distribute the PGs for each pool onto al

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-19 Thread David Turner
pressed_original=0.5? The > second ratio does not look too impressive given the file contents. > > 4) Is there any way to get uncompressed data compressed as a background > task like scrub? > > If you have the time to look at these questions, this would be great. Most > import

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-18 Thread David Turner
>> 2018-10-13 20:58:53.750164 mon.a [WRN] Health check update: Reduced data > >> availability: 3 pgs inactive (PG_AVAILABILITY) > >> ... > >> > >> so there is a timeframe of > 90s whee PGs are inactive and unavail - > >> this would at least ex

Re: [ceph-users] ceph pg/pgp number calculation

2018-10-18 Thread David Turner
Not all pools need the same amount of PGs. When you get to so many pools you want to start calculating how much data each pool will have. If 1 of your pools will have 80% of your data in it, it should have 80% of your PGs. The metadata pools for rgw likely won't need more than 8 or so PGs each. If

Re: [ceph-users] SSD for MON/MGR/MDS

2018-10-15 Thread David Turner
Mgr and MDS do not use physical space on a disk. Mons do use the disk and benefit from SSDs, but they write a lot of stuff all the time. Depending why the SSDs aren't suitable for OSDs, they might not be suitable for mons either. On Mon, Oct 15, 2018, 7:16 AM ST Wong (ITSC) wrote: > Hi all, > >

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-12 Thread David Turner
uot;ceph osd pool set" - > commands, please let me know. > > Best regards, > > = > Frank Schilder > AIT Risø Campus > Bygning 109, rum S14 > > > From: David Turner > Sent: 12 October 2018 15:47:20 > T

Re: [ceph-users] Anyone tested Samsung 860 DCT SSDs?

2018-10-12 Thread David Turner
What do you want to use these for? "5 Year or 0.2 DWPD" is the durability of this drive which is absolutely awful for most every use in Ceph. Possibly if you're using these for data disks (not DB or WAL) and you plan to have a more durable media to host the DB+WAL on... this could work. Or if you

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-12 Thread David Turner
The PGs per OSD does not change unless the OSDs are marked out. You have noout set, so that doesn't change at all during this test. All of your PGs peered quickly at the beginning and then were active+undersized the rest of the time, you never had any blocked requests, and you always had 100MB/s+

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-12 Thread David Turner
provide a source or sample commands? > > Thanks and best regards, > > = > Frank Schilder > AIT Risø Campus > Bygning 109, rum S14 > > ____ > From: David Turner > Sent: 09 October 2018 17:42 > To: Frank Schil

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-11 Thread David Turner
,d83df~1,d83e1~2,d83e5~1,d83e8~1,d83eb~4,d83f0~1,d83f2~1,d83f4~3,d83f8~3,d83fd~2,d8402~1,d8405~1,d8407~1,d840a~2,d840f~1,d8411~1,d8413~3,d8417~3,d841c~4,d8422~4,d8428~2,d842b~1,d842e~1,d8430~1,d8432~5,d843a~1,d843c~3,d8440~5,d8447~1,d844a~1,d844d~1,d844f~1,d8452~1,d8455~1,d8457~1,d8459~2,d845d~2,d84

Re: [ceph-users] Troubleshooting hanging storage backend whenever there is any cluster change

2018-10-11 Thread David Turner
My first guess is to ask what your crush rules are. `ceph osd crush rule dump` along with `ceph osd pool ls detail` would be helpful. Also if you have a `ceph status` output from a time where the VM RBDs aren't working might explain something. On Thu, Oct 11, 2018 at 1:12 PM Nils Fahldieck - Pro

Re: [ceph-users] Inconsistent PG, repair doesn't work

2018-10-11 Thread David Turner
As a part of a repair is queuing a deep scrub. As soon as the repair part is over the deep scrub continues until it is done. On Thu, Oct 11, 2018, 12:26 PM Brett Chancellor wrote: > Does the "repair" function use the same rules as a deep scrub? I couldn't > get one to kick off, until I temporari

Re: [ceph-users] https://ceph-storage.slack.com

2018-10-11 Thread David Turner
I have 4 other slack servers that I'm in for work and personal hobbies. It's just easier for me to maintain one more slack server than have a separate application for IRC. On Thu, Oct 11, 2018, 11:02 AM John Spray wrote: > On Thu, Oct 11, 2018 at 8:44 AM Marc Roos > wrote: > > > > > > Why slack

Re: [ceph-users] OSD log being spammed with BlueStore stupidallocator dump

2018-10-10 Thread David Turner
Not a resolution, but an idea that you've probably thought of. Disabling logging on any affected OSDs (possibly just all of them) seems like a needed step to be able to keep working with this cluster to finish the upgrade and get it healthier. On Wed, Oct 10, 2018 at 6:37 PM Wido den Hollander w

Re: [ceph-users] https://ceph-storage.slack.com

2018-10-10 Thread David Turner
I would like an invite to. drakonst...@gmail.com On Wed, Sep 19, 2018 at 1:02 PM Gregory Farnum wrote: > Done. :) > > On Tue, Sep 18, 2018 at 12:15 PM Alfredo Daniel Rezinovsky < > alfredo.rezinov...@ingenieria.uncuyo.edu.ar> wrote: > >> Can anyone add me to this slack? >> >> with my email alfr

Re: [ceph-users] HEALTH_WARN 2 osd(s) have {NOUP, NODOWN, NOIN, NOOUT} flags set

2018-10-10 Thread David Turner
There is a newer [1] feature to be able to set flags per OSD instead of cluster wide. This way you can prevent a problem host from marking its OSDs down while the rest ofthe cluster is capable of doing so. [2] These commands ought to clear up your status. [1] http://docs.ceph.com/docs/master/rad

Re: [ceph-users] Does anyone use interactive CLI mode?

2018-10-10 Thread David Turner
I know that it existed, but I've never bothered using it. In applications like Python where you can get a different reaction by interacting with it line by line and setting up an environment it is very helpful. Ceph, however, doesn't have any such environment variables that would make this more u

Re: [ceph-users] Can't remove DeleteMarkers in rgw bucket

2018-10-09 Thread David Turner
I would suggest trying to delete the bucket using radosgw-admin. If you can't get that to work, then I would go towards deleting the actual RADOS objects. There are a few threads on the ML that talk about manually deleting a bucket. On Thu, Sep 20, 2018 at 2:04 PM Sean Purdy wrote: > Hi, > > >

Re: [ceph-users] bluestore compression enabled but no data compressed

2018-10-09 Thread David Turner
When I've tested compression before there are 2 places you need to configure compression. On the OSDs in the configuration settings that you mentioned, but also on the [1] pools themselves. If you have the compression mode on the pools set to none, then it doesn't matter what the OSDs configurati

Re: [ceph-users] Error-code 2002/API 405 S3 REST API. Creating a new bucket

2018-10-09 Thread David Turner
Can you outline the process you're using to access the REST API? It's hard to troubleshoot this without knowing how you were trying to do this. On Mon, Sep 17, 2018 at 7:09 PM Michael Schäfer wrote: > Hi, > > We have a problem with the radosgw using the S3 REST API. > Trying to create a new buc

Re: [ceph-users] radosgw bucket stats vs s3cmd du

2018-10-09 Thread David Turner
Have you looked at your Garbage Collection. I would guess that your GC is behind and that radosgw-admin is accounting for that space knowing that it hasn't been freed up yet, whiles 3cmd doesn't see it since it no longer shows in the listing. On Tue, Sep 18, 2018 at 4:45 AM Luis Periquito wrote:

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-10-01 Thread David Turner
> It really seems to be something with RocksDB on centOS. I still think > you can try removing “compression=kNoCompression” from the > filestore_rocksdb_options And/Or check if rocksdb is expecting snappy to be > enabled. > > Thanks, > -Pavan. > > From: David Tur

Re: [ceph-users] mount cephfs from a public network ip of mds

2018-09-30 Thread David Turner
. But other clients with only > 1Gb nic will go through 140.109.0.0 (1Gb LAN) to ask mon or to read/write > to osds. This is why my osds also have 1Gb and 10Gb nics with 140.109.0.0 > and 10.32.0.0 networking respectively. > > Cheers > Joshua > > On Sun, Sep 30, 2018 at 12:09

Re: [ceph-users] mount cephfs from a public network ip of mds

2018-09-29 Thread David Turner
The cluster/private network is only used by the OSDs. Nothing else in ceph or its clients communicate using it. Everything other than osd to osd communication uses the public network. That includes the MONs, MDSs, clients, and anything other than an osd talking to an osd. Nothing else other than os

Re: [ceph-users] Any backfill in our cluster makes the cluster unusable and takes forever

2018-09-27 Thread David Turner
ssion to be enabled, can you try > removing “compression=kNoCompression” from the filestore_rocksdb_options? > And/or you might want to check if rocksdb is expecting snappy to be enabled. > > From: David Turner > Date: Tuesday, September 18, 2018 at 6:01 PM > To: Pavan Rallabhand

Re: [ceph-users] Mimic upgrade failure

2018-09-20 Thread David Turner
> is reporting failure:1 > > I'm working on getting things mostly good again with everything on mimic and > will see if it behaves better. > > Thanks for your input on this David. > > > [global] > mon_initial_members = sephmon1, sephmon2, sephmon3 > mon_host = 10.

  1   2   3   4   5   6   7   8   9   10   >