Hi,
I'm trying cepfs and I have some problems. Here is the context:
All the nodes (in cluster and the clients) are Ubuntu 14.04 with a 3.16
kernel (after apt-get install linux-generic-lts-utopic && reboot).
The cluster:
- one server with just one monitor daemon (RAM 2GB)
- 2 servers (RAM 24GB) with one monitor daemon, ~10 OSDs daemon (one
per disk of 275 GB), and one mds daemon (I use the default
active/standby mode and the pools for cephfs are "data" and "metadata")
The cluster is totally unused (the servers are idle as regards the RAM
and the load overage etc), it's a little cluster for testing, the raw
space is 5172G, number of replicas is 2. Another remark, facing my problem,
I have put in my ceph conf "mds cache size = 1000000" but without lof of
effect (or else I would not be posting this message). Initially, the
cephfs is completely empty.
The clients, "test-cephfs" and "test-cephfs2", have 512MB of RAM. In these
clients, I mount the cephfs like this (with the root account):
~# mkdir /cephfs
~# mount -t ceph 10.0.2.150,10.0.2.151,10.0.2.152:/ /cephfs/ -o
name=cephfs,secretfile=/etc/ceph/ceph.client.cephfs.secret
Then in ceph-testfs, I do:
root@test-cephfs:~# mkdir /cephfs/d1
root@test-cephfs:~# ll /cephfs/
total 4
drwxr-xr-x 1 root root 0 Mar 4 11:45 ./
drwxr-xr-x 24 root root 4096 Mar 4 11:42 ../
drwxr-xr-x 1 root root 0 Mar 4 11:45 d1/
After, in test-cephfs2, I do:
root@test-cephfs2:~# ll /cephfs/
total 4
drwxr-xr-x 1 root root 0 Mar 4 11:45 ./
drwxr-xr-x 24 root root 4096 Mar 4 11:42 ../
drwxrwxrwx 1 root root 0 Mar 4 11:45 d1/
1) Why the unix rights of d1/ are different when I'm in test-cephfs
and when I'm in test-cephfs2? It should be the same, isn't?
2) If I create 100 files in /cephfs/d1/ with test-cephfs:
for i in $(seq 100)
do
echo "$(date +%s.%N)" >/cephfs/d1/f_$i
done
sometimes, in test-cephfs2, when I do a simple:
root@test-cephfs2:~# time \ls -la /cephfs
the command can take 2 or 3 seconds which seems to me very long
for a directory with just 100 files. Generally, if I repeat the
command on test-cephfs2 just after, it's immediate but not always.
I can not reproduce the problem in a determinist way. Sometimes,
to reproduce the problem, I must remove all the files in /cephfs/
on test-cepfs and recreate them. It's very strange. Sometimes and
randomly, something seems to be stalled but I don't know what. I
suspect a problem of mds tuning but, In fact, I don't know what
to do.
Do have an idea of the problem?
3) I plan to use cephfs in production in a project of web servers
(which share together a cephfs storage) but I would like to solve
the issue above before.
If you have any suggestion about cephfs and mds tuning, I am highly
interested.
Thanks in advance for your help.
--
François Lafont
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com