[ceph-users] full osd ssd cluster advise : replication 2x or 3x ?

2014-05-22 Thread Alexandre DERUMIER
Hi, I'm looking to build a full osd ssd cluster, with this config: 6 nodes, each node 10 osd/ ssd drives (dual 10gbit network). (1journal + datas on each osd) ssd drive will be entreprise grade, maybe intel sc3500 800GB (well known ssd) or new Samsung SSD PM853T 960GB (don't have too much

Re: [ceph-users] full osd ssd cluster advise : replication 2x or 3x ?

2014-05-22 Thread Alexandre DERUMIER
dredi 23 Mai 2014 04:57:51 Objet: Re: [ceph-users] full osd ssd cluster advise : replication 2x or 3x ? Hello, On Thu, 22 May 2014 18:00:56 +0200 (CEST) Alexandre DERUMIER wrote: > Hi, > > I'm looking to build a full osd ssd cluster, with this config: > What is your main goa

Re: [ceph-users] full osd ssd cluster advise : replication 2x or 3x ?

2014-05-22 Thread Alexandre DERUMIER
oment) - Mail original - De: "Christian Balzer" À: ceph-users@lists.ceph.com Envoyé: Vendredi 23 Mai 2014 07:29:52 Objet: Re: [ceph-users] full osd ssd cluster advise : replication 2x or 3x ? On Fri, 23 May 2014 07:02:15 +0200 (CEST) Alexandre DERUMIER wrote: > >&g

Re: [ceph-users] full osd ssd cluster advise : replication 2x or 3x ?

2014-05-22 Thread Alexandre DERUMIER
) my main concern, is to known if it's really needed to have replication x3 (mainly for cost price). But I can wait to have lower ssd price next year, and go to 3x if necessary. - Mail original - De: "Alexandre DERUMIER" À: "Christian Balzer" Cc: ceph-use

Re: [ceph-users] collectd / graphite / grafana .. calamari?

2014-05-23 Thread Alexandre DERUMIER
>>https://github.com/rochaporto/collectd-ceph >> >>It has a set of collectd plugins pushing metrics which mostly map what >>the ceph commands return. In the setup we have it pushes them to >>graphite and the displays rely on grafana (check for a screenshot in >>the link above). Thanks for sh

Re: [ceph-users] pgs incomplete; pgs stuck inactive; pgs stuck unclean

2014-05-23 Thread Alexandre DERUMIER
Hi, if you use debian, try to use a recent kernel from backport (>3.10) also check your libleveldb1 version, it should be 1.9.0-1~bpo70+1 (debian wheezy version is too old) I don't see it in ceph repo: http://ceph.com/debian-firefly/pool/main/l/leveldb/ (only for squeeze ~bpo60+1) but you c

Re: [ceph-users] pgs incomplete; pgs stuck inactive; pgs stuck unclean

2014-05-23 Thread Alexandre DERUMIER
De: "jan zeller" À: aderum...@odiso.com Cc: ceph-users@lists.ceph.com Envoyé: Vendredi 23 Mai 2014 13:36:04 Objet: AW: [ceph-users] pgs incomplete; pgs stuck inactive; pgs stuck unclean > -Ursprüngliche Nachricht- > Von: Alexandre DERUMIER [mailto:aderum...@od

Re: [ceph-users] ceph-deploy or manual?

2014-05-27 Thread Alexandre DERUMIER
>> Are people using automation tools like puppet or ansible? http://www.sebastien-han.fr/blog/2014/05/01/vagrant-up-install-ceph-in-one-command/ enjoy ;) - Mail original - De: "Don Talton (dotalton)" À: ceph-users@lists.ceph.com Envoyé: Mardi 27 Mai 2014 18:19:00 Objet: [ceph-use

Re: [ceph-users] Ceph User Committee : welcome Eric Mourgaya

2014-05-29 Thread Alexandre DERUMIER
Congratulations Eric ! - Mail original - De: "Loic Dachary" À: "ceph-users" Envoyé: Jeudi 29 Mai 2014 12:28:56 Objet: [ceph-users] Ceph User Committee : welcome Eric Mourgaya Hi Ceph, Welcome Eric Mourgaya, head of the Ceph User Committee starting May 2014 until the next elect

Re: [ceph-users] Update from dumpling to firefly

2014-05-29 Thread Alexandre DERUMIER
>>crush map has legacy tunables you need t update tunables in the crush, to benefit from last optimisati #ceph osd crush tunables optimal http://ceph.com/docs/master/rados/operations/crush-map/ - Mail original - De: "Fabio - NS3 srl" À: "ceph-users" Envoyé: Jeudi 29 Mai 2014 09

Re: [ceph-users] ceph nodes operanting system suggested

2014-05-29 Thread Alexandre DERUMIER
Hi, I think you can check this wiki: http://ceph.com/docs/master/start/os-recommendations/ currently, only ubuntu 12.04 is deeply tested with inktank (but I think it'll be rhel7 soon ;) the wiki don't have been updated yet for firefly. I known that ceph enterprise users are using dumpling for

Re: [ceph-users] full osd ssd cluster advise : replication 2x or 3x ?

2014-05-29 Thread Alexandre DERUMIER
able to saturated >>10Gbit network. Thanks for your reply ! - Mail original - De: "Łukasz Jagiełło" À: "Alexandre DERUMIER" Cc: ceph-users@lists.ceph.com Envoyé: Mercredi 28 Mai 2014 01:25:40 Objet: Re: [ceph-users] full osd ssd cluster advise : replicati

Re: [ceph-users] Designing a cluster with ceph and benchmark (ceph vs ext4)

2014-06-01 Thread Alexandre DERUMIER
hi! >>See design here: http://adminlinux.com.br/cluster_design.txt >> >># dpkg -l |grep ceph >>ii ceph 0.41-1ubuntu2.1 >>distributed storage >>ii ceph-common0.41-1ubuntu2.1 >>common utilities to mount and

[ceph-users] mellanox SX1012 ethernet|infiniband switch, somebody use it for ceph ?

2014-06-02 Thread Alexandre DERUMIER
se it for ceph ? if yes, what about stability|performance ? with comming rdma support in ceph, it seem to be the perfect solution (and price is very good). Regards, Alexandre Derumier ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists

Re: [ceph-users] mellanox SX1012 ethernet|infiniband switch, somebody use it for ceph ?

2014-06-02 Thread Alexandre DERUMIER
from Ethernet, especially as it relates to network topology. It's cheap and fast, but not always easy to setup if you've got a lot of existing infrastructure. Mark On 06/02/2014 11:29 AM, Alexandre DERUMIER wrote: > Hi, > I'm looking for a fast and cheap 10gbe ethernet

Re: [ceph-users] mellanox SX1012 ethernet|infiniband switch, somebody use it for ceph ?

2014-06-02 Thread Alexandre DERUMIER
. Carlos M. Perez CMP Consulting Services 305-669-1515 > -Original Message- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Alexandre DERUMIER > Sent: Monday, June 2, 2014 12:30 PM > To: ceph-users > Subject: [ceph-users] mellanox

Re: [ceph-users] mellanox SX1012 ethernet|infiniband switch, somebody use it for ceph ?

2014-06-02 Thread Alexandre DERUMIER
I just found this: http://www.mellanox.com/related-docs/whitepapers/WP_Deploying_Ceph_over_High_Performance_Networks.pdf Good to see than ceph begin to be tested by hardware vendor :) Whitepaper include radosbench and fio results - Mail original - De: "Alexandre DERUMIER

Re: [ceph-users] Ceph networks, to bond or not to bond?

2014-06-05 Thread Alexandre DERUMIER
Hi, >>My low-budget setup consists of two gigabit switches, capable of LACP, >>but not stackable. For redundancy, I'd like to have my links spread >>evenly over both switches. If you want to do lacp with both switches, they need to be stackable. (or use active-backup bonding) >>My question wh

Re: [ceph-users] Ceph networks, to bond or not to bond?

2014-06-05 Thread Alexandre DERUMIER
to use both links for public (client-osd). No multipathing like iscsi for example. - Mail original - De: "Sven Budde" À: "Alexandre DERUMIER" Cc: ceph-users@lists.ceph.com Envoyé: Jeudi 5 Juin 2014 18:27:32 Objet: AW: [ceph-users] Ceph networks, to bond or not

[ceph-users] french proxmox meetup

2014-06-05 Thread Alexandre DERUMIER
Hi everybody, we are going to do our first french proxmox in Paris in september http://www.meetup.com/Proxmox-VE-French-Meetup/ And of course, we'll talk about ceph integration in proxmox. So if you are interested, feel free to join us ! Regards, Alexandre ___

Re: [ceph-users] french proxmox meetup

2014-06-06 Thread Alexandre DERUMIER
able. And it's too noisy to discuss anything. Thanks fo the info Loic. For the moment, I have a room in my company building in Paris (for 10-15 peoples), But good to known. I'm waiting for the next ceph meetup too :) - Mail original - De: "Loic Dachary" À: &q

Re: [ceph-users] Minimal io block in rbd

2014-06-10 Thread Alexandre DERUMIER
>>Do for every read 1 Kb rbd will read 4MB from hdd? for write? rados support partial read|write. Note that with erasure code, write need to full rewrite object. (so 4MB) I think that with key-value-store backend (like leveldb), read/write are full too. some interesting notes here : http://eu

[ceph-users] tiering : hit_set_count && hit_set_period memory usage ?

2014-06-11 Thread Alexandre DERUMIER
Hi, I'm reading tiering doc here http://ceph.com/docs/firefly/dev/cache-pool/ " The hit_set_count and hit_set_period define how much time each HitSet should cover, and how many such HitSets to store. Binning accesses over time allows Ceph to independently determine whether an object was accesse

Re: [ceph-users] tiering : hit_set_count && hit_set_period memory usage ?

2014-06-11 Thread Alexandre DERUMIER
, because I have cold datas, but I have full backups jobs running each week, reading all theses cold datas) - Mail original - De: "Gregory Farnum" À: "Alexandre DERUMIER" Cc: "ceph-users" Envoyé: Mercredi 11 Juin 2014 21:56:29 Objet: Re: [ceph-users

Re: [ceph-users] ceph meetup brest

2014-06-16 Thread Alexandre DERUMIER
Hi Eric, timeframe is too short for me, I'm busy this day. I'm interest too by slides if you have them. See you soon, Alexandre - Mail original - De: "eric mourgaya" À: ceph-users@lists.ceph.com, "Loic Dachary" Envoyé: Lundi 16 Juin 2014 08:04:14 Objet: [ceph-users] ceph meetup

Re: [ceph-users] Poor performance on all SSD cluster

2014-06-23 Thread Alexandre DERUMIER
Hi Greg, >>So the only way to improve performance would be to not use O_DIRECT (as this >>should bypass rbd cache as well, right?). yes, indeed O_DIRECT bypass cache. BTW, Do you need to use mysql with O_DIRECT ? default innodb_flush_method is fdatasync, so it should work with cache. (but yo

Re: [ceph-users] Poor performance on all SSD cluster

2014-06-23 Thread Alexandre DERUMIER
I don't known if it's related, but "[Performance] Improvement on DB Performance" http://www.spinics.net/lists/ceph-devel/msg19062.html they are a patch here: https://github.com/ceph/ceph/pull/1848 already pushed in master - Mail original - De: "Robert van Leeuwen" À: ceph-users@l

[ceph-users] ceph.com centos7 repository ?

2014-07-09 Thread Alexandre DERUMIER
Hi, I would like to known if a centos7 respository will be available soon ? Or can I use current rhel7 for the moment ? http://ceph.com/rpm-firefly/rhel7/x86_64/ Cheers, Alexandre ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.c

[ceph-users] inktank-mellanox webinar access ?

2014-07-10 Thread Alexandre DERUMIER
Hi, sorry to spam the mailing list, but they are a inktank mellanox webinar in 10minutes, and I don't have receive access since I have been registered yesterday (same for my co-worker). and the webinar mellanox contact email (conta...@mellanox.com), does not exist Maybe somebody from In

Re: [ceph-users] inktank-mellanox webinar access ?

2014-07-10 Thread Alexandre DERUMIER
Ok, sorry, we have finally receive the login a bit late. Sorry again to have spam the mailing list - Mail original - De: "Alexandre DERUMIER" À: "ceph-users" Envoyé: Jeudi 10 Juillet 2014 16:55:22 Objet: [ceph-users] inktank-mellanox webinar access ? Hi,

Re: [ceph-users] Mixing CEPH versions on new ceph nodes...

2014-07-14 Thread Alexandre DERUMIER
Hi, >>But in reality (yum update or by using ceph-deploy install nodename) - >>the package manager does restart ALL ceph services on that node by its own... debian packages don't restart ceph services on package update, maybe it's a bug in rpm packaging ? - Mail original - De: "

Re: [ceph-users] running Firefly client (0.80.1) against older version (dumpling 0.67.10) cluster?

2014-08-13 Thread Alexandre DERUMIER
Same question here, I'm contributor on proxmox, and we don't known if we can upgrade librbd safely, for users with dumpling cluster. Also, for ceph enterprise , s oes inktank support dumpling enterprise + firefly librbd ? - Mail original - De: "Nigel Williams" À: ceph-users@lis

Re: [ceph-users] Is it safe to enable rbd cache with qemu?

2014-08-22 Thread Alexandre DERUMIER
Hi, for RHEL5, I'm not sure be barriers supported is maybe not implemented in virtio devices,lvm,dm raid and some filesystem, depend of the kernel version. Not sure what is backported in rhel5 kernel see http://monolight.cc/2011/06/barriers-caches-filesystems/ - Mail original -

Re: [ceph-users] Is it safe to enable rbd cache with qemu?

2014-08-22 Thread Alexandre DERUMIER
rivers-windows/issues/3 - Mail original - De: "Yufang" À: "Alexandre DERUMIER" Cc: ceph-users@lists.ceph.com Envoyé: Vendredi 22 Août 2014 18:05:32 Objet: Re: [ceph-users] Is it safe to enable rbd cache with qemu? Thanks, Alexandre. But what about Windows? Does NT

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-08-31 Thread Alexandre DERUMIER
>>Allegedly this model ssd (128G m550) can do 75K 4k random write IOPS >>(running fio on the filesystem I've seen 70K IOPS so is reasonably >>believable). So anyway we are not getting anywhere near the max IOPS >>from our devices. Hi, Just check this: http://www.anandtech.com/show/7864/crucia

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-02 Thread Alexandre DERUMIER
Hi Sebastien, >>I got 6340 IOPS on a single OSD SSD. (journal and data on the same >>partition). Shouldn't it better to have 2 partitions, 1 for journal and 1 for datas ? (I'm thinking about filesystem write syncs) - Mail original - De: "Sebastien Han" À: "Somnath Roy" Cc: ce

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-02 Thread Alexandre DERUMIER
Do you have same results, if you launch 2 fio benchs in parallel on 2 differents rbd volumes ? - Mail original - De: "Sebastien Han" À: "Cédric Lemarchand" Cc: "Alexandre DERUMIER" , ceph-users@lists.ceph.com Envoyé: Mardi 2 Septembre 2014 13

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-02 Thread Alexandre DERUMIER
De: "Sebastien Han" À: "Alexandre DERUMIER" Cc: ceph-users@lists.ceph.com, "Cédric Lemarchand" Envoyé: Mardi 2 Septembre 2014 15:25:05 Objet: Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS Well the last time I ran two processes

Re: [ceph-users] Ceph Day Paris Schedule Posted

2014-09-04 Thread Alexandre DERUMIER
I was waiting for the schedule, topics seem to be interesting. I'm going to register now :) BTW, are the speeches in french or english? (As I see loic,sebastian and yann as speakers) - Mail original - De: "Patrick McGarry" À: "Ceph Devel" , "Ceph-User" , "Ceph User Committee" Envoyé

Re: [ceph-users] resizing the OSD

2014-09-05 Thread Alexandre DERUMIER
>>Is there a way to resize the OSD without bringing the cluster down? What is the HEALTH state of your cluster ? If it's OK, simply replace the osd disk by a bigger one ? - Mail original - De: "JIten Shah" À: ceph-us...@ceph.com Envoyé: Samedi 6 Septembre 2014 00:31:01 Objet: [c

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-10 Thread Alexandre DERUMIER
Hi Sebastien, here my first results with crucial m550 (I'll send result with intel s3500 later): - 3 nodes - dell r620 without expander backplane - sas controller : lsi LSI 9207 (no hardware raid or cache) - 2 x E5-2603v2 1.8GHz (4cores) - 32GB ram - network : 2xgigabit link lacp + 2xgigabit lac

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-11 Thread Alexandre DERUMIER
> Le 11/09/2014 08:20, Alexandre DERUMIER a écrit : >> Hi Sebastien, >> >> here my first results with crucial m550 (I'll send result with intel s3500 >> later): >> >> - 3 nodes >> - dell r620 without expander backplane >> - sas cont

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-11 Thread Alexandre DERUMIER
ename=/dev/sdb --direct=1 --rw=write --bs=4k --numjobs=2 --group_reporting --invalidate=0 --name=ab --sync=1 bw=177575KB/s, iops=44393 - Mail original - De: "Alexandre DERUMIER" À: "Cedric Lemarchand" Cc: ceph-users@lists.ceph.com Envoyé: Vendredi 12 Sept

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-11 Thread Alexandre DERUMIER
/s wrqm/s r/s w/s rkB/s wkB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sdb 0,00 29,00 0,00 3075,00 0,00 36748,50 23,90 0,29 0,10 0,00 0,10 0,05 15,20 So, the write bottleneck seem to be in ceph. I will send s3500 result today - Mail original - De: "Alexandre DER

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-12 Thread Alexandre DERUMIER
it w_await svctm %util sdb 0,00 1563,000,00 9880,00 0,00 75223,5015,23 2,090,210,000,21 0,07 80,00 - Mail original ----- De: "Alexandre DERUMIER" À: "Cedric Lemarchand" Cc: ceph-users@lists.ceph.com Envoyé: Vendredi 12 Sept

Re: [ceph-users] OpTracker optimization

2014-09-13 Thread Alexandre DERUMIER
Hi, as ceph user, It could be wonderfull to have it for Giant, optracker performance impact is really huge (See my ssd benchmark on ceph user mailing) Regards, Alexandre Derumier - Mail original - De: "Somnath Roy" À: "Samuel Just" Cc: "Sage Weil"

[ceph-users] best libleveldb version ?

2014-09-15 Thread Alexandre DERUMIER
Hi, I would like to known with libleveldb should be us with firefly. I'm using debian wheezy which provide really old libleveldb (I don't use it), and in wheezy backport 1.17 is provided. But in intank repositories , I see that 1.9 is provide for some distribs. So, what is the best/tested ve

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-16 Thread Alexandre DERUMIER
hursday, could be great if we could share experience about ceph and ssd. Alexandre. - Mail original - De: "Sebastien Han" À: "Alexandre DERUMIER" Cc: ceph-users@lists.ceph.com Envoyé: Mardi 16 Septembre 2014 15:32:59 Objet: Re: [ceph-users] [Single OSD pe

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-17 Thread Alexandre DERUMIER
my ssds model for my production cluster (target 2015),I'll have a look for this optimus drives - Mail original - De: "Somnath Roy" À: "Mark Kirkwood" , "Alexandre DERUMIER" , "Sebastien Han" Cc: ceph-users@lists.ceph.com Envoyé: Mer

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-18 Thread Alexandre DERUMIER
ly or Giant ? I'll do benchs with 6 osd dc3500 tomorrow to compare firefly and giant. - Mail original - De: "Jian Zhang" À: "Sebastien Han" , "Alexandre DERUMIER" Cc: ceph-users@lists.ceph.com Envoyé: Jeudi 18 Septembre 2014 08:12:32 Objet: RE:

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-19 Thread Alexandre DERUMIER
_op_threads = 5 filestore_op_threads = 4 bw=62094KB/s, iops=15523 giant with same tuning --- bw=247073KB/s, iops=61768 ! I think I could reach more, but my 2 gigabit link are satured. - Mail original ----- De: "Alexandre DERUMIER" À: "Ji

Re: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS

2014-09-19 Thread Alexandre DERUMIER
B/s, iops=29532 - Mail original - De: "Jian Zhang" À: "Alexandre DERUMIER" Cc: ceph-users@lists.ceph.com Envoyé: Vendredi 19 Septembre 2014 10:21:38 Objet: RE: [ceph-users] [Single OSD performance on SSD] Can't go over 3, 2K IOPS Thanks for this gre

Re: [ceph-users] erasure code : number of chunks for a small cluster ?

2015-02-02 Thread Alexandre DERUMIER
ing :) Thanks Loic - Mail original - De: "Loic Dachary" À: "aderumier" , "ceph-users" Envoyé: Dimanche 1 Février 2015 18:42:51 Objet: Re: [ceph-users] erasure code : number of chunks for a small cluster ? Hi Alexandre, On 01/02/2015 18:15, Alexandre DERUMIER wr

Re: [ceph-users] erasure code : number of chunks for a small cluster ?

2015-02-02 Thread Alexandre DERUMIER
y sense. I start with 5 hosts and use k:3, m:2. In this case two hdds can fail or one host can be down for maintenance. Udo PS: you also can't change k+m on a pool later... On 01.02.2015 18:15, Alexandre DERUMIER wrote: > Hi, > > I'm currently trying to understand how to se

Re: [ceph-users] features of the next stable release

2015-02-03 Thread Alexandre DERUMIER
Hi, From my tests with giant, this was the cpu which limit the performance on osd. I'm going to do some benchmark with 2x10 cores 3,1ghz for 6ssd next month. I'll post results on the mailing list. - Mail original - De: "mad Engineer" À: "Gregory Farnum" Cc: "ceph-users" Envoyé: Ma

Re: [ceph-users] Update 0.80.7 to 0.80.8 -- Restart Order

2015-02-03 Thread Alexandre DERUMIER
debian deb packages update are not restarting services. (So, I think it should be the same for ubuntu). you need to restart daemons in this order: -monitor -osd -mds -rados gateway http://ceph.com/docs/master/install/upgrading-ceph/ - Mail original - De: "Stephen Jahl" À: "Gregory Fa

Re: [ceph-users] ceph Performance random write is more then sequential

2015-02-04 Thread Alexandre DERUMIER
Hi, >>What I saw after enabling RBD cache it is working as expected, means >>sequential write has better MBps than random write. can somebody explain this >>behaviour ? This is because rbd_cache merge coalesced ios in bigger ios, so it's working only with sequential workload. you'll do less i

Re: [ceph-users] erasure code : number of chunks for a small cluster ?

2015-02-06 Thread Alexandre DERUMIER
Oh, I didn't thinked about this. Thanks Hector ! - Mail original - De: "Hector Martin" À: "ceph-users" Envoyé: Vendredi 6 Février 2015 09:06:29 Objet: Re: [ceph-users] erasure code : number of chunks for a small cluster ? On 02/02/15 03:38, Udo Lembke wrote: > With 3 hosts only you c

[ceph-users] crush tunables : optimal : upgrade from firefly to hammer behaviour ?

2015-02-08 Thread Alexandre DERUMIER
Hi, I'm currently use crush tunables "optimal" value. If I upgrade from firefly to hammer, does the optimal value will upgrade to optimal values for hammer. So, does my clients (qemu-librbd) need to be also upgraded to hammer to support new hammer features ? If yes, I think to: - change c

Re: [ceph-users] crush tunables : optimal : upgrade from firefly to hammer behaviour ?

2015-02-08 Thread Alexandre DERUMIER
m firefly to hammer behaviour ? On Mon, 9 Feb 2015, Alexandre DERUMIER wrote: > Hi, > > I'm currently use crush tunables "optimal" value. > > If I upgrade from firefly to hammer, does the optimal value will upgrade > to optimal values for hammer. The tunable

Re: [ceph-users] crush tunables : optimal : upgrade from firefly to hammer behaviour ?

2015-02-08 Thread Alexandre DERUMIER
hours of benchmarking. So be careful. crush optimal sets it to the current optimal value. So you need to upgrade your clients first and then set it to optimal again. Stefan Excuse my typo s ent from my mobile phone. Am 09.02.2015 um 06:32 schrieb Alexandre DERUMIER < aderum...@odiso.c

Re: [ceph-users] stuck with dell perc 710p / (aka mega raid 2208?)

2015-02-10 Thread Alexandre DERUMIER
Hi, you need to import foreign config from openmanage webui. somewhere in storage controller BTW, I'm currently testing new dell r630 with a perc h330 ( lsi 3008) With this controller, it's possible to do hardware for some disks, and passthrough for some others disks. So, perfect for ceph :)

[ceph-users] mongodb on top of rbd volumes (through krbd) ?

2015-02-11 Thread Alexandre DERUMIER
Hi, I'm currently running a big mongodb cluster, around 2TB, (sharding + replication). And I have a lot of problems with mongo replication (out of syncs and need to full replicate again and again datas between my mongo replicats). So, I thinked to use rbd to replicate the storage and keep onl

Re: [ceph-users] mongodb on top of rbd volumes (through krbd) ?

2015-02-12 Thread Alexandre DERUMIER
the behavior of mongo when a shard is unavailable for some reason (crash or network partition) ? If shard3 is on the wrong side of a network partition and uses RBD, it will hang. Is it something that mongo will gracefully handle ? I have no experience in this but I'm curious about this

Re: [ceph-users] re: Upgrade 0.80.5 to 0.80.8 --the VM's read requestbecome too slow

2015-02-12 Thread Alexandre DERUMIER
>>Hi, >>Can you test with disabling rbd_cache ? >>I remember of a bug detected in giant, not sure it's also the case for fireflt This was this tracker: http://tracker.ceph.com/issues/9513 But It has been solved and backported to firefly. Also, can you test 0.80.6 and 0.80.7 ? - Mai

Re: [ceph-users] Upgrade 0.80.5 to 0.80.8 --the VM's read request become too slow

2015-02-12 Thread Alexandre DERUMIER
Hi, Do you have also tested 0.80.6 and 0.80.7 librbd ? could be usefull to search commits in git. (I'm not sure that all changes are in the release note) - Mail original - De: "杨万元" À: "ceph-users" Envoyé: Jeudi 12 Février 2015 04:14:15 Objet: [ceph-users] Upgrade 0.80.5 to 0.80.8 --t

Re: [ceph-users] OSD slow requests causing disk aborts in KVM

2015-02-12 Thread Alexandre DERUMIER
>>To my surprise however these slow requests caused aborts from the block >>device on the VM side, which ended up corrupting files This is very strange, you shouldn't have corruption. Do you use writeback ? if yes, do you have disable barrier on your filesystem ? (What is the qemu version ? gue

Re: [ceph-users] OSD slow requests causing disk aborts in KVM

2015-02-13 Thread Alexandre DERUMIER
>>Can this timeout be increased in some way? I've searched around and found the >>/sys/block/sdx/device/timeout knob, which in my case is set to 30s. yes, sure echo 60 > /sys/block/sdx/device/timeout for 60s for example - Mail original - De: "Krzysztof Nowicki" À: "Andrey Korolyov"

Re: [ceph-users] Introducing "Learning Ceph" : The First ever Book on Ceph

2015-02-13 Thread Alexandre DERUMIER
Just buy it. Nice book, I don't have read yet all the book, but it seem to cover all ceph features. Good job ! - Mail original - De: "Karan Singh" À: "Ceph Community" , "ceph-users" , ceph-maintain...@ceph.com, "ceph-users" , "ceph-devel" Cc: "Sage Weil" , "don" Envoyé: Jeudi 5 Fé

Re: [ceph-users] Ceph Dumpling/Firefly/Hammer SSD/Memstore performance comparison

2015-02-18 Thread Alexandre DERUMIER
Nice Work Mark ! I don't see any tuning about sharding in the config file sample (osd_op_num_threads_per_shard,osd_op_num_shards,...) as you only use 1 ssd for the bench, I think it should improve results for hammer ? - Mail original - De: "Mark Nelson" À: "ceph-devel" Cc: "ceph-us

Re: [ceph-users] Ceph Dumpling/Firefly/Hammer SSD/Memstore performance comparison

2015-02-18 Thread Alexandre DERUMIER
ll. Have you done any testing here? It could be an interesting follow-up paper. Mark On 02/18/2015 02:34 AM, Alexandre DERUMIER wrote: > Nice Work Mark ! > > I don't see any tuning about sharding in the config file sample > > (osd_op_num_threads_per_shard,osd_op_num_s

[ceph-users] new ssd intel s3610, has somebody tested them ?

2015-02-19 Thread Alexandre DERUMIER
Hi, Intel has just released new ssd s3610: http://www.anandtech.com/show/8954/intel-launches-ssd-dc-s3610-s3710-enterprise-ssds endurance is 10x bigger than 3500, for 10% cost addition. Has somebody already tested them ? Regards, Alexandre ___ ceph-

Re: [ceph-users] re: Upgrade 0.80.5 to 0.80.8 --the VM's read requestbecome too slow

2015-02-21 Thread Alexandre DERUMIER
his problem is cause from 0.80.8 2015-02-12 19:33 GMT+08:00 Alexandre DERUMIER < aderum...@odiso.com > : >>Hi, >>Can you test with disabling rbd_cache ? >>I remember of a bug detected in giant, not sure it's also the case for >>fireflt This was this tra

Re: [ceph-users] Ceph 0.87-1

2015-02-25 Thread Alexandre DERUMIER
Hi, I known that Loic Dachary was currently working on backporting new feature on giant, I see that 0.87.1 has been tagged in git too: here the difference: https://github.com/ceph/ceph/compare/v0.87...v0.87.1 Loic, any annoucement/release note, yet ? - Mail original - De: "Lindsay M

Re: [ceph-users] v0.87.1 Giant released

2015-02-26 Thread Alexandre DERUMIER
I just upgraded my debian giant cluster, 1)on each node: - apt-get update apt-get dist-upgrade 2)on each node: - /etc/init.d/ceph restart mon #ceph -w ---> verify that HEALTH is ok before doing another node 3)on each node: - /etc/init.d/ceph restart osd #ce

Re: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results

2015-02-26 Thread Alexandre DERUMIER
l - De: "Mark Nelson" À: "ceph-devel" , "ceph-users" Envoyé: Jeudi 26 Février 2015 05:44:15 Objet: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results Hi Everyone, In the Ceph Dumpling/Firefly/Hammer SSD/Memstore performance comparison thread, Alexandre D

Re: [ceph-users] Possibly misleading/outdated documentation about qemu/kvm and rbd cache settings

2015-02-27 Thread Alexandre DERUMIER
Hi, from qemu rbd.c if (flags & BDRV_O_NOCACHE) { rados_conf_set(s->cluster, "rbd_cache", "false"); } else { rados_conf_set(s->cluster, "rbd_cache", "true"); } and block.c int bdrv_parse_cache_flags(const char *mode, int *flags) { *flags &= ~BDRV_O_CACHE_MASK;

Re: [ceph-users] Extreme slowness in SSD cluster with 3 nodes and 9 OSD with 3.16-3 kernel

2015-02-28 Thread Alexandre DERUMIER
Hi, First, test if your ssd can write fast with O_DSYNC check this blog: http://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-device/ Then, try with ceph Giant (or maybe wait for Hammer), because they are a lot of optimisations for ssd for threads sh

Re: [ceph-users] Extreme slowness in SSD cluster with 3 nodes and 9 OSD with 3.16-3 kernel

2015-02-28 Thread Alexandre DERUMIER
and 9 OSD with 3.16-3 kernel > Am 28.02.2015 um 12:43 schrieb Alexandre DERUMIER : > > Hi, > > First, test if your ssd can write fast with O_DSYNC > check this blog: > http://www.sebastien-han.fr/blog/2014/10/10/ceph-how-to-test-if-your-ssd-is-suitable-as-a-journal-dev

Re: [ceph-users] Extreme slowness in SSD cluster with 3 nodes and 9 OSD with 3.16-3 kernel

2015-02-28 Thread Alexandre DERUMIER
As optimisation, try to set ioscheduler to noop, and also enable rbd_cache=true. (It's really helping for for sequential writes) but your results seem quite low, 926kb/s with 4k, it's only 200io/s. check if you don't have any big network latencies, or mtu fragementation problem. Maybe also t

Re: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results

2015-03-01 Thread Alexandre DERUMIER
compare - Mail original - De: "Mark Nelson" À: "ceph-devel" , "ceph-users" Envoyé: Jeudi 26 Février 2015 05:44:15 Objet: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results Hi Everyone, In the Ceph Dumpling/Firefly/Hammer SSD/Memstore performa

Re: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results

2015-03-01 Thread Alexandre DERUMIER
Ceph Hammer OSD Shard Tuning Test Results Can I ask what xio and simple messenger are and the differences? Kind regards Kevin Walker +968 9765 1742 On 1 Mar 2015, at 18:38, Alexandre DERUMIER wrote: Hi Mark, I found an previous bench from Vu Pham (it's was about simplemessenger vs

Re: [ceph-users] Permanente Mount RBD blocs device RHEL7

2015-03-02 Thread Alexandre DERUMIER
Hi, maybe this can help you: http://www.sebastien-han.fr/blog/2013/11/22/map-slash-unmap-rbd-device-on-boot-slash-shutdown/ Regards, Alexandre - Mail original - De: "Jesus Chavez (jeschave)" À: "ceph-users" Envoyé: Lundi 2 Mars 2015 11:14:49 Objet: [ceph-users] Permanente Mount RBD

Re: [ceph-users] Ceph Hammer OSD Shard Tuning Test Results

2015-03-02 Thread Alexandre DERUMIER
tics match the benchmark result numbers? I also did some tests on fdcache, though just glancing at the results it doesn't look like tweaking those parameters had much effect. Mark On 03/01/2015 08:38 AM, Alexandre DERUMIER wrote: > Hi Mark, > > I found an previous bench from

Re: [ceph-users] Update 0.80.5 to 0.80.8 --the VM's read request become too slow

2015-03-02 Thread Alexandre DERUMIER
I think this will be fixed in next firefly point release tracker for firefly 0.80.8 speed decrease http://tracker.ceph.com/issues/10956 Jason Dillaman link it to the famous object_cacher bug: http://tracker.ceph.com/issues/9854 - Mail original - De: "Gregory Farnum" À: "Nathan O'Sulliv

Re: [ceph-users] Perf problem after upgrade from dumpling to firefly

2015-03-04 Thread Alexandre DERUMIER
Hi, maybe this is related ?: http://tracker.ceph.com/issues/9503 "Dumpling: removing many snapshots in a short time makes OSDs go berserk" http://tracker.ceph.com/issues/9487 "dumpling: snaptrimmer causes slow requests while backfilling. osd_snap_trim_sleep not helping" http://lists.opennebula

Re: [ceph-users] Perf problem after upgrade from dumpling to firefly

2015-03-04 Thread Alexandre DERUMIER
HDD since firefly. And yes, the problem hang the production at night during snap trimming. I suppose there is a new OSD parameter which change behavior of the journal, or something like that. But didn't find anything about that. Olivier Le mercredi 04 mars 2015 à 14:44 +0100, Alexandre D

Re: [ceph-users] Perf problem after upgrade from dumpling to firefly

2015-03-04 Thread Alexandre DERUMIER
Le mercredi 04 mars 2015 à 15:05 +0100, Alexandre DERUMIER a écrit : > >>The load problem is permanent : I have twice IO/s on HDD since firefly. > > Oh, permanent, that's strange. (If you don't see more traffic coming from > clients, I don't understand...) &

Re: [ceph-users] Perf problem after upgrade from dumpling to firefly

2015-03-04 Thread Alexandre DERUMIER
Objet: Re: [ceph-users] Perf problem after upgrade from dumpling to firefly Only writes ;) Le mercredi 04 mars 2015 à 16:19 +0100, Alexandre DERUMIER a écrit : > >>The change is only on OSD (and not on OSD journal). > > do you see twice iops for read and write ? > >

Re: [ceph-users] CEPH hardware recommendations and cluster design questions

2015-03-04 Thread Alexandre DERUMIER
Hi for hardware, inktank have good guides here: http://www.inktank.com/resource/inktank-hardware-selection-guide/ http://www.inktank.com/resource/inktank-hardware-configuration-guide/ ceph works well with multiple osd daemon (1 osd by disk), so you should not use raid. (xfs is the recommended fs

Re: [ceph-users] Strange krbd behaviour with queue depths

2015-03-06 Thread Alexandre DERUMIER
Hi, do you have tried with differents io schedulers to compare ? - Mail original - De: "Nick Fisk" À: "ceph-users" Envoyé: Jeudi 5 Mars 2015 18:17:27 Objet: [ceph-users] Strange krbd behaviour with queue depths I’m seeing a strange queue depth behaviour with a kernel mapped RBD, libr

Re: [ceph-users] Multiple OSD's in a Each node with replica 2

2015-03-06 Thread Alexandre DERUMIER
>>Is it possible all replicas of an object to be saved in the same node? No. (until you don't wrongly modify the crushmap manually) >>Is it possible to lose any? with replicat x2, if you loose 2osd on 2differents nodes, with the same object inside, you'll lost the object >>Is there a mechanis

Re: [ceph-users] Strange krbd behaviour with queue depths

2015-03-06 Thread Alexandre DERUMIER
want to try that. But, I think it is available in the latest kernel >> only. >> Librbd is running with TCP_NODELAY = true by default, you may want to try >> with ms_tcp_nodelay = false to simulate the similar behavior with librbd. >> >> Thanks & Regards >> Somnath

Re: [ceph-users] Changing pg_num => RBD VM down !

2015-03-14 Thread Alexandre DERUMIER
>>And at this moment, some of the VM stored on this pool were stopped (on >>some hosts, not all, it depends, no logic) do you use librbd or krbd for theses vm ? Is the guest os crashed ? or the qemu process killed?(which seem really strange) - Mail original - De: "Florent Bautista"

Re: [ceph-users] [SPAM] Changing pg_num => RBD VM down !

2015-03-16 Thread Alexandre DERUMIER
>>That full system slows down, OK, but brutal stop... This is strange, that could be: - qemu crash, maybe a bug in rbd block storage (if you use librbd) - oom-killer on you host (any logs ?) what is your qemu version ? - Mail original - De: "Florent Bautista" À: "ceph-users" Envoyé:

Re: [ceph-users] [SPAM] Changing pg_num => RBD VM down !

2015-03-16 Thread Alexandre DERUMIER
ent Bautista" À: "aderumier" Cc: "ceph-users" Envoyé: Lundi 16 Mars 2015 11:14:45 Objet: Re: [ceph-users] [SPAM] Changing pg_num => RBD VM down ! On 03/16/2015 11:03 AM, Alexandre DERUMIER wrote: > This is strange, that could be: > > - qemu crash, maybe

Re: [ceph-users] [SPAM] Changing pg_num => RBD VM down !

2015-03-16 Thread Alexandre DERUMIER
derumier" Cc: "ceph-users" Envoyé: Lundi 16 Mars 2015 12:35:11 Objet: Re: [ceph-users] [SPAM] Changing pg_num => RBD VM down ! On 03/16/2015 12:23 PM, Alexandre DERUMIER wrote: >>> We use Proxmox, so I think it uses librbd ? > As It's me that I made t

Re: [ceph-users] SSD Hardware recommendation

2015-03-18 Thread Alexandre DERUMIER
Hi Josef, I'm going to benchmark a 3nodes cluster with 6ssd each node (2x10 cores 3,1ghz). From my previous bench, you need fast cpus if you need a lot of iops, and writes are lot more expansive than reads. Now i'm you are doing only small iops (big blocks / big throughput), you don't need too

Re: [ceph-users] RBD read-ahead not working in 0.87.1

2015-03-18 Thread Alexandre DERUMIER
Hi, I don't known how rbd read-ahead is working, but with qemu virtio-scsi, you can have read merge request (for sequential reads), so it's doing bigger ops to ceph cluster and improve throughput. virtio-blk merge request will be supported in coming qemu 2.3. (I'm not sure of virtio-win driver

Re: [ceph-users] SSD Hardware recommendation

2015-03-18 Thread Alexandre DERUMIER
’ve been done with it before I ordered :) I will most likely order this week so I will see it when the stuff is being assembled :o Do you feel that there something in the setup that could be better if you would decide on hardware as of today? Also, will you try out replication set of 2 as well?

  1   2   3   4   5   >