Dear,
  we are studying the possibility to migrate our FS in the next year to
cephfs. I know that it is not prepare for production environments yet, but
we are planning to play with it in the next months deploying a basic
testbed.
  Reading the documentation, I see 3 mons, 1 mds and several ods's (both in
physical machines..I have understood). Is this true?
  On the other hand I do not understand the fail-over mechanism for clients
when have mounted a FS, looking at documentation :

   *ceph-fuse* [ -m *monaddr*:*port* ] *mountpoint* [ *fuse options* ]
You have to specify (hardcode) the "*monaddr*:

*port", if this mon (ip) is down, what happen, Do you lost the fs on that
node?, or there is a generic dns-rrd implementation for mons?? *

*  Is there any implementation for "tiering" or "HSM" at software level, I
mean, can I mix different type of disk (ssds and SATA) on diferent pools,
and migrate data between them automatically (most used, size, last time
access)  *



*Please could anyone clarify to me this point?*

*Regards, I*

-- 
####################################
Iban Cabrillo Bartolome
Instituto de Fisica de Cantabria (IFCA)
Santander, Spain
Tel: +34942200969
####################################
Bertrand Russell:
*"El problema con el mundo es que los estúpidos están seguros de todo y los
inteligentes están llenos de dudas*"
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to