I am on Cent OS 6.5 and i am using: [root@node1 ~]# rpm -qa | grep gluster glusterfs-rdma-3.4.2-1.el6.x86_64 glusterfs-server-3.4.2-1.el6.x86_64 glusterfs-fuse-3.4.2-1.el6.x86_64 glusterfs-libs-3.4.2-1.el6.x86_64 glusterfs-3.4.2-1.el6.x86_64 glusterfs-api-3.4.2-1.el6.x86_64 glusterfs-cli-3.4.2-1.el6.x86_64 vdsm-gluster-4.13.3-3.el6.noarch
[root@node1 ~]# uname -a Linux node1.hq.imos.net 2.6.32-431.el6.x86_64 #1 SMP Fri Nov 22 03:15:09 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux Thanks, Mario On Sat, Feb 8, 2014 at 10:06 PM, Samuli Heinonen <[email protected]>wrote: > Hello, > > What version of GlusterFS you are using? > > ml ml <[email protected]> kirjoitti 8.2.2014 kello 21.24: > > anyone? > > On Friday, February 7, 2014, ml ml <[email protected]> wrote: > >> Hello List, >> >> i set up a Cluster with 2 Nodes and Glusterfs. >> >> >> gluster> volume info all >> >> Volume Name: Repl2 >> Type: Replicate >> Volume ID: 8af9b282-8b60-4d71-a0fd-9116b8fdcca7 >> Status: Started >> Number of Bricks: 1 x 2 = 2 >> Transport-type: tcp >> Bricks: >> Brick1: node1.local:/data >> Brick2: node2.local:/data >> Options Reconfigured: >> auth.allow: * >> user.cifs: enable >> nfs.disable: off >> >> >> I turned node2 off. Just to make sure i have not network bottle neck and >> that it will not replicate for my first benchmarks. >> >> >> My first test with bonnie on my local raw disk of node1 gave me 130MB/sec >> write speed. >> Then i did the same test on my cluster dir /data: 130MB/sec >> Then i did the write test in a freshly installed Debian 7 vm: 10MB/sec >> >> This is terrible and i wonder why?! >> >> My tests where made with: >> bonnie++ -u root -s <double mem) -d <dir> >> >> Here are my bonnie results: http://oi62.tinypic.com/20aara0.jpg >> >> >> Since node2 is turned off, this cant be a network bottle neck. >> >> Any ideas? >> >> Thanks, >> Mario >> > _______________________________________________ > Users mailing list > [email protected] > http://lists.ovirt.org/mailman/listinfo/users > > >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

