Hello Charles 

Need some more clarification with your setup , Did you mean 

1) There is 1 SSD ( 60 GB ) on each server i.e 6 SSD on all 6 servers ? 

2) your osd.3 , osd.4 , osd.5 uses same journal ( /dev/sdf2 ) ? 

Regards 
Karan Singh 

----- Original Message -----

From: "charles L" <charlesboy...@hotmail.com> 
To: "ceph dev" <majord...@vger.kernel.org>, ceph-us...@ceph.com 
Sent: Thursday, 31 October, 2013 6:24:13 AM 
Subject: [ceph-users] testing ceph 

Hi, 
Pls is this a good setup for a production environment test of ceph? My focus is 
on the SSD ... should it be partitioned(sdf1,2 ,3,4) and shared by the four 
OSDs on a host? or is this a better configuration for the SSD to be just one 
partition(sdf1) while all osd uses that one partition? 
my setup: 
- 6 Servers with one 250gb boot disk for OS(sda), 
four-2Tb Disks each for the OSDs i.e Total disks = 6x4 = 24 disks (sdb -sde) 
and one-60GB SSD for Osd Journal(sdf). 
-RAM = 32GB on each server with 2 GB network link. 
hostname for servers: Server1 -Server6 

[osd.0] 
host = server1 
devs = /dev/sdb 
osd journal = /dev/sdf1 
[osd.1] 
host = server1 
devs = /dev/sdc 
osd journal = /dev/sdf2 

[osd.3] 
host = server1 
devs = /dev/sdd 
osd journal = /dev/sdf2 

[osd.4] 
host = server1 
devs = /dev/sde 
osd journal = /dev/sdf2 
[osd.5] 
host = server2 
devs = /dev/sdb 
osd journal = /dev/sdf2 
... 
[osd.23] 
host = server6 
devs = /dev/sde 
osd journal = /dev/sdf2 

Thanks. 

_______________________________________________ 
ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to