Hello,

On Fri, 11 Apr 2014 09:48:56 +0800 Punit Dambiwal wrote:

> Hi,
> 
> What is the drawback to run the journals on the RAID1...??
>
Did you read what I wrote below?
 
> My plan is 2 SSD RAID1 (then i will create virtual disks for OS as well
> as for every OSD).That means one virtual disk of OS and other 24 virtual
> disks for journals.... ??
>
Assuming your SSD (even though you have 2 since you're using RAID1 you only
have the write capacity of one) is a decent Intel DC 3700 400GB, the
maximum write speed of that is 460MB/s. 
Meaning 2 things:
a) if your're using 10GbE for networking, you already limited your max
throughput to that, instead of the 1GB/s your network could theoretically
achieve.
b) at that speed your SSD journal can serve about 4 actual OSD disks at
full speed. You crippled the performance potential of your storage node
immensely. 

Even with the limitation of the single min-SAS link to the backplane in
this storage server (it is a 4x lane, so 24Gb/s total) you would need 4 or
5 dedicated SSDs at the above speed to saturate that link.

Of course the bandwidth of that link is theoretical, with 24 disks to
server it will be lower in practice. 

> Please suggest me better way to do this ??
>
I already did below.
 
Regards,

Christian
> 
> On Wed, Apr 9, 2014 at 4:02 PM, Christian Balzer <ch...@gol.com> wrote:
> 
> > On Wed, 9 Apr 2014 14:59:30 +0800 Punit Dambiwal wrote:
> >
> > > Hi,
> > >
> > > I have changed my plan and now i want to use the following supermicro
> > > server :-
> > >
> > > SuperStorage Server 6047R-E1R24L
> > >
> > > Can any one tell me....is this server is good for the OSD
> > > nodes...two ssd on RAID1 (OS & journal) and 24 HDD for OSD (JBOD on
> > > the motherboard controller).
> > >
> > Wrong on so many levels.
> >
> > Firstly, 2 SSDs (really just one if you're using RAID1 for the journal
> > partitions as well) for 24 OSDs.
> > The SSD will be a speed bottleneck and also have to handle ALL the
> > writes that ever happen to the whole machine (meaning it will wear out
> > quickly).
> >
> > If you want/need SSD journals, a sensible ratio would be 3-4 OSDs per
> > journal SSD (partitioned into the respective amount of journals).
> > So something like 6 SSDs and 18 HDDs.
> >
> > Secondly that backplane is connected to the HBA with one mini-SAS link.
> > That means at best 4 lanes of 6Gb/s for 24 drives, but it might be just
> > one lane, the manual is typical Supermicro quality. =.=
> > Another, potentially massive bottleneck.
> >
> > Also what are your goals in term of throughput, IOPS here?
> > If you're planning on getting lots of 24 disk boxes, fine.
> > Otherwise you might be better off getting smaller nodes.
> >
> > Regards,
> >
> > Christian.
> > >
> > >
> > >
> > >
> > >
> > > On Fri, Apr 4, 2014 at 11:51 AM, Ирек Фасихов <malm...@gmail.com>
> > > wrote:
> > >
> > > > You need to use Dell OpenManage:
> > > >
> > > > https://linux.dell.com/repo/hardware/.
> > > >
> > > >
> > > >
> > > > 2014-04-04 7:26 GMT+04:00 Punit Dambiwal <hypu...@gmail.com>:
> > > >
> > > >> Hi,
> > > >>
> > > >> I want to use Dell R515/R510 for the OSD node purpose....
> > > >>
> > > >> 1. 2*SSD for OS purpose (Raid 1)
> > > >> 2. 10* Segate 3.5' HDDx 3TB for OSD purpose (No RAID...JBOD)
> > > >>
> > > >> To create JBOD...i created all 10 HDD as raid0....but the problem
> > > >> is when i will plug out the HDD from the server and plug-in
> > > >> again,i need to import the RAID configuration again to make this
> > > >> OSD working....
> > > >>
> > > >> Can anyone suggest me good way to do this ??
> > > >>
> > > >> Thanks,
> > > >> Punit
> > > >>
> > > >> _______________________________________________
> > > >> ceph-users mailing list
> > > >> ceph-users@lists.ceph.com
> > > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > >>
> > > >>
> > > >
> > > >
> > > > --
> > > > С уважением, Фасихов Ирек Нургаязович
> > > > Моб.: +79229045757
> > > >
> >
> >
> > --
> > Christian Balzer        Network/Systems Engineer
> > ch...@gol.com           Global OnLine Japan/Fusion Communications
> > http://www.gol.com/
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >


-- 
Christian Balzer        Network/Systems Engineer                
ch...@gol.com           Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to