[ceph-users] jewel 10.2.5 cephfs fsync write issue

2017-01-18 Thread jiajia zhong
hi guys:

I am using kernel client (4.9) to mount cephfs (10.2.5) that just upgraded
from Hammer(0.94.9).

It became slow while doing fsync write  VS  same process in Hammer.
(Yes, I am sure, fsync is the key.)


1. anyone knows what's goning on ?
2. any way to import that?

===
btw, actually, It's the same slow if I deploy a new cluster, with kernel
client 4.2 or fuse client. ~
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Issue with upgrade from 0.94.9 to 10.2.5

2017-01-18 Thread Piotr Dałek

On 01/17/2017 12:52 PM, Piotr Dałek wrote:

During our testing we found out that during upgrade from 0.94.9 to 10.2.5
we're hitting issue http://tracker.ceph.com/issues/17386 ("Upgrading 0.94.6
-> 0.94.9 saturating mon node networking"). Apparently, there's a few
commits for both hammer and jewel which are supposed to fix this issue for
upgrades from 0.94.6 to 0.94.9 (and possibly for others), but we're still
seeing this upgrading to Jewel, and symptoms are exactly same - after
upgrading MONs, each not yet upgraded OSD takes full OSDMap from monitors
after failing the CRC check. Anyone else encountered this?


http://tracker.ceph.com/issues/18582

--
Piotr Dałek
piotr.da...@corp.ovh.com
https://www.ovh.com/us/
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] failing to respond to capability release, mds cache size?

2017-01-18 Thread Gregory Farnum
On Tue, Jan 17, 2017 at 10:56 AM, Darrell Enns  wrote:
> Thanks for the info, I'll be sure to "dump_ops_in_flight" and "session ls" if 
> it crops up again. Is there any other info you can think of that might be 
> useful? I want to make sure I capture all the evidence needed if it happens 
> again.

If you can dump the MDS cache that might turn out to be useful too;
otherwise there's probably not much you can get unless your kernel
includes dynamic debug and you can play with the
/sys/kernel/debug/ceph that is described at
http://docs.ceph.com/docs/master/cephfs/troubleshooting/#kernel-mount-debugging
But I'll leave that up to Zheng or somebody who works on the kernel client.
-Greg
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Ceph uses more raw space than expected

2017-01-18 Thread Pavel Shub
Hi all,

I'm running a 6 node 24 OSD cluster, Jewel 10.2.5 with kernel 4.8.

I put about 1TB of data in the cluster, with all pools having size 3. Yet
about 5TB of raw disk is used as opposed to the expected 3TB.

result of ceph -s:

  pgmap v1057361: 2400 pgs, 3 pools, 984 GB data, 125 Mobjects
5039 GB used, 12353 GB / 17393 GB avail
2398 active+clean
   2 active+clean+scrubbing

Result of ceph df:

GLOBAL:
SIZE   AVAIL  RAW USED %RAW USED
17393G 12353G5039G 28.97
POOLS:
NAMEID USED   %USED MAX AVAIL OBJECTS
rbd 1801G 17.12 3880G206299
cephfs_data 2182G  4.49 3880G 130875410
cephfs_metadata 3  32540k 0 3880G201555

Result of ceph osd dump:

pool 1 'rbd' replicated size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 800 pgp_num 800 last_change 482 flags hashpspool
stripe_width 0
pool 2 'cephfs_data' replicated size 3 min_size 1 crush_ruleset 0
object_hash rjenkins pg_num 800 pgp_num 800 last_change 410 flags
hashpspool crash_replay_interval 45 stripe_width 0
pool 3 'cephfs_metadata' replicated size 3 min_size 1 crush_ruleset 0
object_hash rjenkins pg_num 800 pgp_num 800 last_change 408 flags
hashpspool stripe_width 0

The cluster was set up with ceph-deploy. So each OSD drive is formatted
with xfs. One more thing I should mention is that I'm using the
experimental directory fragmentation feature in cephfs.
After scouring the mailing list I've found this other post that seems to be
related or the same issue:
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-October/004901.html

Does anyone know if this is a bug or is it legitimate overhead that I
failed to account for?

Thanks,
Pavel
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] RadosGW Performance on Copy

2017-01-18 Thread Eric Choi
Performance related data I found so far on RadosGW don't contain anything
on Copy operation.  Can anyone comment on the Copy operation? Is it
basically just a GET + INSERT?

Thanks,

Eric
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] GSOC 2017 Submissions Open Tomorrow

2017-01-18 Thread Patrick McGarry
Hey cephers,

Just wanted to send a friendly reminder that the Google Summer of Code
program has their submissions window opening tomorrow, and will remain
open for 3 weeks. If you are interested in mentoring or suggesting a
project please contact me as soon as possible. Thanks.


-- 

Best Regards,

Patrick McGarry
Director Ceph Community || Red Hat
http://ceph.com  ||  http://community.redhat.com
@scuttlemonkey || @ceph
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] rgw static website docs 404

2017-01-18 Thread Ben Hines
Are there docs on the RGW static website feature?

I found 'rgw enable static website' config setting only via the mailing
list. A search for 'static' on ceph.com turns up release notes, but no
other documentation. Anyone have pointers on how to set this up and what i
can do with it? Does it require using dns based buckets, for example? I'd
like to be able to hit a website with http:,
ideally. (without the browser forcing it to download)

thanks,

-Ben
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] rgw static website docs 404

2017-01-18 Thread Ben Hines
Aha! Found some docs here in the RHCS site:

https://access.redhat.com/documentation/en/red-hat-ceph-storage/2/paged/object-gateway-guide-for-red-hat-enterprise-linux/chapter-2-configuration

Really, ceph.com should have all this too...

-Ben

On Wed, Jan 18, 2017 at 5:15 PM, Ben Hines  wrote:

> Are there docs on the RGW static website feature?
>
> I found 'rgw enable static website' config setting only via the mailing
> list. A search for 'static' on ceph.com turns up release notes, but no
> other documentation. Anyone have pointers on how to set this up and what i
> can do with it? Does it require using dns based buckets, for example? I'd
> like to be able to hit a website with http:,
> ideally. (without the browser forcing it to download)
>
> thanks,
>
> -Ben
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Does this indicate a "CPU bottleneck"?

2017-01-18 Thread 许雪寒
Hi, everyone.

Recently, we did some stress test on ceph using three machines. We tested the 
IOPS of the whole small cluster when there are 1~8 OSDs per machines separately 
and the result is as follows:

 OSD num per machine fio iops

1   
10k

2   
16.5k

3   
22k

4   
23.5k

5   
26k

6   
27k

7   
27k

8   
28k

As shown above, it seems that there is some kind of bottleneck when there are 
more than 4 OSDs per machine. Meanwhile, we observed that the CPU %idle during 
the test, shown below, has also some kind of correlation with the number of 
OSDs per machine.

 OSD num per machine CPU idle

1   
74%

2   
52%

3   
30%

4   
25%

5   
24%

6   
17%

7   
14%

8   
11%

It seems that with the number of OSDs per machine increasing, the CPU idle time 
is reducing and the reduce rate Is also decreasing, can we come to the 
conclusion that CPU is the performance bottleneck in this test?

Thank you:)
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com