Re: [ceph-users] Mon stuck in synchronizing after upgrading from Hammer to Jewel

2017-07-04 Thread jiajia zhong
refer to http://docs.ceph.com/docs/master/rados/operations/add-or-rm-mons/ I recalled we encoutered the same issue after upgrading to Jewel :(. 2017-07-05 11:21 GMT+08:00 许雪寒 : > Hi, everyone. > > Recently, we upgraded one of clusters from Hammer to Jewel. However, after > upgrading one of our m

Re: [ceph-users] is there any way to speed up cache evicting?

2017-06-02 Thread jiajia zhong
get_full_ratio only concerned the clean objects (evict), any misunderstanding ? This fixed our cluster speeds during times when the cache pool was cleaning > up. The problem we hypothesized was that it was the prices of choosing what > to clean vs what to keep was causing. > > On Fri

Re: [ceph-users] is there any way to speed up cache evicting?

2017-06-02 Thread jiajia zhong
thank you for your guide :), It's making sense. 2017-06-02 16:17 GMT+08:00 Christian Balzer : > > Hello, > > On Fri, 2 Jun 2017 14:30:56 +0800 jiajia zhong wrote: > > > christian, thanks for your reply. > > > > 2017-06-02 11:39 GMT+08:00 Christian Balzer :

Re: [ceph-users] is there any way to speed up cache evicting?

2017-06-01 Thread jiajia zhong
christian, thanks for your reply. 2017-06-02 11:39 GMT+08:00 Christian Balzer : > On Fri, 2 Jun 2017 10:30:46 +0800 jiajia zhong wrote: > > > hi guys: > > > > Our ceph cluster is working with tier cache. > If so, then I suppose you read all the discussions here

[ceph-users] is there any way to speed up cache evicting?

2017-06-01 Thread jiajia zhong
hi guys: Our ceph cluster is working with tier cache. I am running "rados -p data_cache cache-try-flush-evict-all" to evict all the objects. But It a bit slow 1. Is there any way to speed up the evicting? 2. Is evicting triggered by itself good enough for cluster ? 3. Does the flushing and evic

Re: [ceph-users] Mix HDDs and SSDs togheter

2017-03-05 Thread jiajia zhong
we are using mixed too, intel PCIE 400G SSD * 8 for metadata pool and tier caching pool for our cephfs. *plus:* *'osd crush update on start = false*' as Vladimir replied. 2017-03-03 20:33 GMT+08:00 Дробышевский, Владимир : > Hi, Matteo! > > Yes, I'm using mixed cluster in production but it's

Re: [ceph-users] ceph upgrade from hammer to jewel

2017-02-23 Thread jiajia zhong
are you sure you have ceph-fuse upgraded? #ceph-fuse --version 2017-02-23 16:07 GMT+08:00 gjprabu : > Hi Team, > > We upgraded ceph version from 0.94.9 hammer to 10.2.5 jewel . > Still some clients are showing older version while mounting with debug > mode, is this caused any issue w

Re: [ceph-users] Migrate cephfs metadata to SSD in running cluster

2017-02-20 Thread jiajia zhong
n? > > I am kind of a little paranoid to start playing around with crush rules in > the running system. > > Regards, > > Mike > > On 1/5/17 11:40 PM, jiajia zhong wrote: > >> >> >> 2017-01-04 23:52 GMT+08:00 Mike Miller > millermike...@gmail.com&g

Re: [ceph-users] 2 of 3 monitors down and to recover

2017-02-09 Thread jiajia zhong
hi taotao :) you can follow http://docs.ceph.com/docs/master/rados/operations/add-or-rm-mons/ , Remove the non-surviving or problematic monitors. For example ... remember to backup the surviving monitor data before any futher. 2017-02-10 9:52 GMT+08:00 何涛涛(云平台事业部) : > Hey, > > I tried to sim

Re: [ceph-users] machine hangs & soft lockups with 10.2.2 / kernel 4.4.0

2017-01-23 Thread jiajia zhong
try a newer kernel. like 4.8 2017-01-24 0:37 GMT+08:00 Matthew Vernon : > Hi, > > We have a 9-node ceph cluster, running 10.2.2 and kernel 4.4.0 (Ubuntu > Xenial). We're seeing both machines freezing (nothing in logs on the > machine, which is entirely unresponsive to anything except the power >

[ceph-users] jewel 10.2.5 cephfs fsync write issue

2017-01-18 Thread jiajia zhong
hi guys: I am using kernel client (4.9) to mount cephfs (10.2.5) that just upgraded from Hammer(0.94.9). It became slow while doing fsync write VS same process in Hammer. (Yes, I am sure, fsync is the key.) 1. anyone knows what's goning on ? 2. any way to import that? ===

Re: [ceph-users] 答复: Pipe "deadlock" in Hammer, 0.94.5

2017-01-12 Thread jiajia zhong
g on the socket by “Pipe::tcp_read_wait”. So, I think, when Pipe::do_recv > is called, it shouldn’t get an EAGAIN, which means it wouldn’t act as > blocked. Is this so? > This really confuses me. > > > 发件人: jiajia zhong [mailto:zhong2p...@gmail.com] > 发送时间: 2017年1月12日 18:22 > 收

Re: [ceph-users] Pipe "deadlock" in Hammer, 0.94.5

2017-01-12 Thread jiajia zhong
if errno is EAGAIN for recv, the Pipe:do_recv just acts as blocked. so 2017-01-12 16:34 GMT+08:00 许雪寒 : > Hi, everyone. > > Recently, we did some experiment to test the stability of the ceph > cluster. We used Hammer version which is the mostly used version of online > cluster. One of the scenari

Re: [ceph-users] Write back cache removal

2017-01-10 Thread jiajia zhong
It's fixed since v0.94.6, http://ceph.com/releases/v0-94-6-hammer-released/ - fs: CephFS restriction on removing cache tiers is overly strict ( issue#11504 , pr#6402 , John Spray) but you have to make sure you

Re: [ceph-users] Migrate cephfs metadata to SSD in running cluster

2017-01-05 Thread jiajia zhong
2017-01-04 23:52 GMT+08:00 Mike Miller : > Wido, all, > > can you point me to the "recent benchmarks" so I can have a look? > How do you define "performance"? I would not expect cephFS throughput to > change, but it is surprising to me that metadata on SSD will have no > measurable effect on laten

Re: [ceph-users] Analysing ceph performance with SSD journal, 10gbe NIC and 2 replicas -Hammer release

2017-01-05 Thread jiajia zhong
2017-01-06 11:10 GMT+08:00 kevin parrikar : > Hello All, > > I have setup a ceph cluster based on 0.94.6 release in 2 servers each > with 80Gb intel s3510 and 2x3 Tb 7.2 SATA disks,16 CPU,24G RAM > which is connected to a 10G switch with a replica of 2 [ i will add 3 more > servers to the cluster

Re: [ceph-users] Ceph - Health and Monitoring

2017-01-04 Thread jiajia zhong
actually, what you need is an ceph-common package (ubuntu) which contains /usr/bin/ceph, You have to be sure the command's going to be executed on which host. make sure the keys and ceph.conf are correctly configured on that host. you could just run the commands to make sure the configure's ok. eg

Re: [ceph-users] 2 OSD's per drive , unable to start the osd's

2016-12-16 Thread JiaJia Zhong
In your scenario, don't use ceph-disk follow http://docs.ceph.com/docs/jewel/rados/operations/add-or-rm-osds/ -- Original -- From: "Burkhard Linke"; Date: Fri, Dec 16, 2016 05:09 PM To: "CEPH list"; Subject: Re: [ceph-users] 2 OSD's per drive , unable to st

Re: [ceph-users] can cache-mode be set to readproxy for tiercachewith ceph 0.94.9 ?

2016-12-16 Thread JiaJia Zhong
hi skinjo, forgot to ask that if it's necessary to disconnect all the client before doing set-overlay ? we didn't sweep the clients out while setting overlay -- Original -- From: "JiaJia Zhong"; Date: Wed, Dec 14, 2016 11:24 AM To:

Re: [ceph-users] can cache-mode be set to readproxy for tiercachewith ceph 0.94.9 ?

2016-12-13 Thread JiaJia Zhong
-- Original -- From: "Shinobu Kinjo"; Date: Wed, Dec 14, 2016 10:56 AM To: "JiaJia Zhong"; Cc: "CEPH list"; "ukernel"; Subject: Re: [ceph-users] can cache-mode be set to readproxy for tiercachewith ceph 0.94.9

Re: [ceph-users] can cache-mode be set to readproxy for tier cachewith ceph 0.94.9 ?

2016-12-13 Thread JiaJia Zhong
ds = false -- Original -- From: "Shinobu Kinjo"; Date: Tue, Dec 13, 2016 06:21 PM To: "JiaJia Zhong"; Cc: "CEPH list"; "ukernel"; Subject: Re: [ceph-users] can cache-mode be set to readproxy for tier cachewith ceph 0.9

[ceph-users] can cache-mode be set to readproxy for tier cache with ceph 0.94.9 ?

2016-12-12 Thread JiaJia Zhong
hi cephers: we are using ceph hammer 0.94.9, yes, It's not the latest ( jewel), with some ssd osds for tiering, cache-mode is set to readproxy, everything seems to be as expected, but when reading some small files from cephfs, we got 0 bytes. I did some search and got the be

Re: [ceph-users] Ceph performance is too good (impossible..)...

2016-12-11 Thread JiaJia Zhong
>> 3. After the test, in a.txt, we got bw=1162.7MB/s, in b.txt, we get >> bw=3579.6MB/s. mostly, due to your kernel buffer of client host -- Original -- From: "Somnath Roy"; Date: Mon, Dec 12, 2016 09:47 AM To: "V Plus"; "CEPH list"; Subject: Re: [ceph

Re: [ceph-users] No module named rados

2016-11-28 Thread JiaJia Zhong
hi, since you are playing on centos7, why not following http://docs.ceph.com/docs/master/install/get-packages/ or just downloading the binary packages form https://download.ceph.com/rpm-jewel/ ? :) if you insist to install ceph from ceph-10.2.2.tar.gz, please follow http://docs.ceph.com/docs/gi

Re: [ceph-users] undefined symbol: rados_inconsistent_pg_list

2016-11-28 Thread JiaJia Zhong
hi, 1.try below, remove /root/.python-eggs/rados-0-py2.7-linux-x86_64.egg-tmp/ if you are sure that you want to keep it, back it up. 2. the command you ran, #cp -vf /usr/local/lib/python2.7/site-packages/* /usr/lib64/python2.7/ , generally, this is not recommended. -- Orig

[ceph-users] about using SSD in cephfs, attached with some quantified benchmarks

2016-11-25 Thread JiaJia Zhong
confusing questions: (ceph0.94) 1. Is there any way to cache the whole metadata datas into MDS's memory ? (metadata osds dates-async> MDS memory) I dunno if I misunderstand the role of mds :(, so many post threads that advising Using SSD osds for metadata. the metadata stores Inod

[ceph-users] how to get the default CRUSH map that should be generated by ceph itself ?

2016-11-23 Thread JiaJia Zhong
hi, folksIs there any way I could get the original(default) crush map after some manual mofifications? eg: assuming no one had tuned the crush map, the map state named ORIGINAL , which was maintained by ceph itself. if I modified the crush map, map state named COMMIT1,

Re: [ceph-users] cephfs (rbd) read performance low - where is thebottleneck?

2016-11-23 Thread JiaJia Zhong
Mike, if you run mount.ceph with "-v" options, you may get "ceph: Unknown mount option rsize", actually, you could ignore this, the rsize and rasize will both be passed to mount syscall. I belive that you have had the cephfs mounted successfully, run "mount" in terminal to check the actual moun