Re: [ceph-users] limited disk slots - should I ran OS on SD card ?

2018-08-14 Thread Burkhard Linke
Hi, AFAIk SD cards (and SATA DOMs) do not have any kind of wear-leveling support. Even if the crappy write endurance of these storage systems would be enough to operate a server for several years on average, you will always have some hot spots with higher than usual write activity. This is t

Re: [ceph-users] cephfs kernel client hangs

2018-08-14 Thread Yan, Zheng
On Mon, Aug 13, 2018 at 9:55 PM Zhenshi Zhou wrote: > > Hi Burkhard, > I'm sure the user has permission to read and write. Besides, we're not using > EC data pools. > Now the situation is that any openration to a specific file, the command will > hang. > Operations to any other files won't hang.

Re: [ceph-users] cephfs kernel client hangs

2018-08-14 Thread Zhenshi Zhou
kernel client Yan, Zheng 于2018年8月14日周二 下午3:13写道: > On Mon, Aug 13, 2018 at 9:55 PM Zhenshi Zhou wrote: > > > > Hi Burkhard, > > I'm sure the user has permission to read and write. Besides, we're not > using EC data pools. > > Now the situation is that any openration to a specific file, the comm

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Glen Baars
Hello Jason, I can confirm that your tests work on our cluster with a newly created image. We still can’t get the current images to use a different object pool. Do you think that maybe another feature is incompatible with this feature? Below is a log of the issue. :~# rbd info RBD_HDD/2ef34a9

[ceph-users] Ceph upgrade Jewel to Luminous

2018-08-14 Thread Jaime Ibar
Hi all, we're running Ceph Jewel 10.2.10 in our cluster and we plan to upgrade to latest Luminous release(12.2.7). Jewel 10.2.11 was released one month ago and ours plans were upgrade to this release first and then upgrade to Luminous, but as someone reported osd's crashes after upgradin

[ceph-users] Ceph upgrade Jewel to Luminous

2018-08-14 Thread Jaime Ibar
Hi all, we're running Ceph Jewel 10.2.10 in our cluster and we plan to upgrade to latest Luminous release(12.2.7). Jewel 10.2.11 was released one month ago and ours plans were upgrade to this release first and then upgrade to Luminous, but as someone reported osd's crashes after upgrading to

[ceph-users] mimic/bluestore cluster can't allocate space for bluefs

2018-08-14 Thread Jakub Stańczak
Hello All! I am using mimic full bluestore cluster with pure RGW workload. We use AWS i3 instance family for osd machines - each instance has 1 NVMe disk which is split into 4 partitions and each of those partitions is devoted to bluestore block device. We use 1 device per partition - so everythin

Re: [ceph-users] Replicating between two datacenters without decompiling CRUSH map

2018-08-14 Thread Paul Emmerich
IIRC this will create a rule that tries to selects n independent data centers Check the actual generated rule to validate this. I think the only way to express "3 copies across two data centers" is by explicitly using the two data centers in the rule as in: (pseudo code) take dc1 chooseleaf 1 typ

Re: [ceph-users] Stale PG data loss

2018-08-14 Thread Surya Bala
We have configured cephfs . Yes i can restore the data. But i need to know which files are corrupted. So that i can delete those files and copy them again. In normal state i can get inode id of files and mapping inode id with object id. So that i can get files to object mapping . using ceph osd ma

Re: [ceph-users] Inconsistent PG could not be repaired

2018-08-14 Thread Arvydas Opulskis
Thanks for suggestion about restarting OSD's, but this doesn't work either. Anyway, I managed to fix second unrepairing PG by getting object from OSD and saving it again via rados, but still no luck with first one. I think, I found main problem why this doesn't work. Seems, object is not overwritt

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Jason Dillaman
On Tue, Aug 14, 2018 at 4:08 AM Glen Baars wrote: > Hello Jason, > > > > I can confirm that your tests work on our cluster with a newly created > image. > > > > We still can’t get the current images to use a different object pool. Do > you think that maybe another feature is incompatible with thi

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Glen Baars
Hello Jason, I will also complete testing of a few combinations tomorrow to try and isolate the issue now that we can get it to work with a new image. The cluster started out at 12.2.3 bluestore so there shouldn’t be any old issues from previous versions. Kind regards, Glen Baars From: Jason D

Re: [ceph-users] bad crc/signature errors

2018-08-14 Thread Ilya Dryomov
On Mon, Aug 13, 2018 at 5:57 PM Nikola Ciprich wrote: > > Hi Ilya, > > hmm, OK, I'm not sure now whether this is the bug which I'm > experiencing.. I've had read_partial_message / bad crc/signature > problem occurance on the second cluster in short period even though > we're on the same ceph ver

Re: [ceph-users] bad crc/signature errors

2018-08-14 Thread Nikola Ciprich
> > Hi Ilya, > > > > hmm, OK, I'm not sure now whether this is the bug which I'm > > experiencing.. I've had read_partial_message / bad crc/signature > > problem occurance on the second cluster in short period even though > > we're on the same ceph version (12.2.5) for quite long time (almost sin

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Jason Dillaman
I tried w/ a rbd CLI from 12.2.7 and I still don't have an issue enabling journaling on a different pool: $ rbd info rbd/foo rbd image 'foo': size 1024 MB in 256 objects order 22 (4096 kB objects) block_name_prefix: rbd_data.101e6b8b4567 format: 2 features: layering, exclusive-lock, object-map, fa

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Glen Baars
Hello Jason, I have tried with and without ‘rbd journal pool = rbd’ in the ceph.conf. it doesn’t seem to make a difference. Also, here is the output: rbd image-meta list RBD-HDD/2ef34a96-27e0-4ae7-9888-fd33c38f657a There are 0 metadata on this image. Kind regards, Glen Baars From: Jason Dillam

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Jason Dillaman
On Tue, Aug 14, 2018 at 9:19 AM Glen Baars wrote: > Hello Jason, > > > > I have tried with and without ‘rbd journal pool = rbd’ in the ceph.conf. > it doesn’t seem to make a difference. > It should be SSDPOOL, but regardless, I am at a loss as to why it's not working for you. You can try appendi

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Glen Baars
Hello Jason, I have now narrowed it down. If the image has an exclusive lock – the journal doesn’t go on the correct pool. Kind regards, Glen Baars From: Jason Dillaman Sent: Tuesday, 14 August 2018 9:29 PM To: Glen Baars Cc: ceph-users Subject: Re: [ceph-users] RBD journal feature On Tue,

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Jason Dillaman
On Tue, Aug 14, 2018 at 9:31 AM Glen Baars wrote: > Hello Jason, > > > > I have now narrowed it down. > > > > If the image has an exclusive lock – the journal doesn’t go on the correct > pool. > OK, that makes sense. If you have an active client on the image holding the lock, the request to enab

Re: [ceph-users] RBD journal feature

2018-08-14 Thread Glen Baars
Hello Jason, Thanks for your help. Here is the output you asked for also. https://pastebin.com/dKH6mpwk Kind regards, Glen Baars From: Jason Dillaman Sent: Tuesday, 14 August 2018 9:33 PM To: Glen Baars Cc: ceph-users Subject: Re: [ceph-users] RBD journal feature On Tue, Aug 14, 2018 at 9:31

Re: [ceph-users] Slow rbd reads (fast writes) with luminous + bluestore

2018-08-14 Thread Emmanuel Lacour
Le 13/08/2018 à 16:58, Jason Dillaman a écrit : > > See [1] for ways to tweak the bluestore cache sizes. I believe that by > default, bluestore will not cache any data but instead will only > attempt to cache its key/value store and metadata. I suppose too because default ratio is to cache as much

[ceph-users] rhel/centos7 spectre meltdown experience

2018-08-14 Thread Marc Roos
Did anyone notice any performance loss on osd, mon, rgw nodes because of the spectre/meltdown updates? What is general practice concerning these updates? Sort of follow up on this discussion. https://www.mail-archive.com/ceph-users@lists.ceph.com/msg43136.html https://access.redhat.com/arti

Re: [ceph-users] mimic/bluestore cluster can't allocate space for bluefs

2018-08-14 Thread Igor Fedotov
Hi Jakub, for the crashing OSD could you please set debug_bluestore=10 bluestore_bluefs_balance_failure_dump_interval=1 and collect more logs. This will hopefully provide more insight on why additional space isn't allocated for bluefs. Thanks, Igor On 8/14/2018 12:41 PM, Jakub Stańczak

Re: [ceph-users] Ceph upgrade Jewel to Luminous

2018-08-14 Thread Thomas White
Hi Jaime, Upgrading directly should not be a problem. It is usually recommended to go to the latest minor release before upgrading major versions, but my own migration from 10.2.10 to 12.2.5 went seamlessly and I can’t see of any technical limitation which would hinder or prevent this proces

Re: [ceph-users] Inconsistent PG could not be repaired

2018-08-14 Thread Thomas White
Hi Arvydas, The error seems to suggest this is not an issue with your object data, but the expected object digest data. I am unable to access where I stored my very hacky diagnosis process for this, but our eventual fix was to locate the bucket or files affected and then rename an object wit

Re: [ceph-users] limited disk slots - should I ran OS on SD card ?

2018-08-14 Thread Paul Emmerich
I've seen the OS running on SATA DOMs and cheap USB sticks. It works well for some time, and then it just falls apart. Paul 2018-08-14 9:12 GMT+02:00 Burkhard Linke : > Hi, > > > AFAIk SD cards (and SATA DOMs) do not have any kind of wear-leveling > support. Even if the crappy write endurance of

[ceph-users] BlueStore wal vs. db size

2018-08-14 Thread Robert Stanford
I am keeping the wal and db for a ceph cluster on an SSD. I am using the masif_bluestore_block_db_size / masif_bluestore_block_wal_size parameters in ceph.conf to specify how big they should be. Should these values be the same, or should one be much larger than the other? R

Re: [ceph-users] BlueStore wal vs. db size

2018-08-14 Thread Wido den Hollander
On 08/15/2018 04:17 AM, Robert Stanford wrote: > I am keeping the wal and db for a ceph cluster on an SSD.  I am using > the masif_bluestore_block_db_size / masif_bluestore_block_wal_size > parameters in ceph.conf to specify how big they should be.  Should these > values be the same, or should on