confusing questions: (ceph0.94)

1. Is there any way to cache the whole metadata datas into MDS's memory ?   
(metadata osds dates-----async----> MDS memory)


I dunno if I misunderstand the role of mds :(,   so many post threads that 
advising Using SSD osds for metadata.
the metadata stores Inode information for files. Yes, It's fast to stat, ls, 
readdir for cephfs, 
but If metadatas could be cached in memory,  metadata osds datas 
-----async----> MDS memory, I guess, this may be better ?
we can use ssd journals, so write speed would not be the bottleneck.  cached 
mdatadatas are not large even if there are huge number of files.  ( I got that 
MooseFS strores all metadata in memory ?)


2. 
Any descriptions for the Journal under hood ? though it's like swap partition 
for Linux ~


using a Intel PCIE SSD as journals of HDD osds, 
I ran command blow for a  rough benchmark all of the osds simultaneously,


# for i in $(ps aux | grep osd | awk '{print $14}' | grep -v "^$" | sort); do 
ceph tell osd.$i bench & done


compared to anther HOST without SSD journal, these got a better bytes_per_sec, 
rising about more than 100%.
HDD journal OSDS  30MB/s ------> HDD OSDS with SSD journal more than 60MB/s.   
(12 osds/host, hosts are almost same)


MB/s (HDD journal + HDD)
39
35
35
35
33
31
29
26
26
26
26
25

the top 39MB/s one an stata SSD OSD with stata SSD journal, but the speed seems 
to be  not faster than the otheres with HDD Journal + HDD Data.


MB/s (PCIE SSD Journal + HDD)
195
129
92
88
71
71
65
61
57
54
52
50

that 195MB/s is PCIE SSD Journal + SSD Data, the speed seems to be very fast. 
others are PCIE SSD Journal + HDD Data,




"bytes_per_sec": 166451390.000000 for single bench on (PCIE Journal + HDD)    
158.74MB/s

"bytes_per_sec": 78472933.000000 for single bench on   (HDD Journal + HDD)      
 74.83MB/s


It seems that  "data ---> HDD Journal" is probable the main bottleneck ? how to 
track this
data ----> SSD Journal ------> osd data partition
data ----> HDD Journal -----> osd data partition


3.
any  cache or memory suggestion for better performance for cephfs?


key ceph.conf as below
[global]
osd pool default size = 2
osd pool default min size = 1
osd pool default pg num = 512
osd pool default pgp num = 512
osd journal size = 10000


[mds]
mds cache size = 11474836


[osd]
osd op threads = 4
filestore op threads = 4
osd crush update on start = false
#256M
osd max write size = 256
#256M
journal max write bytes = 268435456
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to