No, I've never seen this issue on the Fedora stock kernels.
So either my workflow is not triggering it on the Fedora software stack
or the issues is CentOS / RHEL - specific.
Anyway I will file the ceph-fuse bug then.
On Tue, Feb 2, 2016 at 12:43 PM, Yan, Zheng <uker...@gmail.com> wrote:
On Tue, Feb 2, 2016 at 5:32 PM, Mykola Dvornik
<mykola.dvor...@gmail.com> wrote:
One of my clients is using
4.3.5-300.fc23.x86_64 (Fedora release 23)
did you encounter this problem on client using 4.3.5 kernel? If you
did, this issue should be ceph-fuse bug.
while all the other clients reply on
3.10.0-327.4.4.el7.x86_64 (CentOS Linux release 7.2.1511)
Should I file report a bug on the RedHat bugzilla?
you can open a bug at http://tracker.ceph.com/projects/cephfs/issues
Regards
Yan, Zheng
On Tue, Feb 2, 2016 at 8:57 AM, Yan, Zheng <uker...@gmail.com>
wrote:
On Tue, Feb 2, 2016 at 2:27 AM, Mykola Dvornik
<mykola.dvor...@gmail.com>
wrote:
What version are you running on your servers and clients?
Are you using 4.1 or 4.2 kernel?
https://bugzilla.kernel.org/show_bug.cgi?id=104911. Upgrade to 4.3+
kernel
or 4.1.17 kernel or 4.2.8 kernel can resolve this issue.
On the clients: ceph-fuse --version ceph version 9.2.0
(bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) MDS/OSD/MON: ceph
--version ceph
version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299) Exactly
what
changes are you making that aren't visible? I am creating some new
files in
non-root folders. What's the output of "ceph -s"? ceph -s cluster
98d72518-6619-4b5c-b148-9a781ef13bcb health HEALTH_OK monmap e1: 1
mons at
{000-s-ragnarok=XXX.XXX.XXX.XXX:6789/0} election epoch 1, quorum 0
000-s-ragnarok mdsmap e576: 1/1/1 up {0=000-s-ragnarok=up:active}
osdmap
e233: 16 osds: 16 up, 16 in flags sortbitwise pgmap v1927636: 1088
pgs, 2
pools, 1907 GB data, 2428 kobjects 3844 GB used, 25949 GB / 29793
GB avail
1088 active+clean client io 4381 B/s wr, 2 op In addition on the
clients'
side I have cat /etc/fuse.conf user_allow_other auto_cache
large_read
max_write = 16777216 max_read = 16777216 -Mykola On Mon, Feb 1,
2016 at 5:06
PM, Gregory Farnum <gfar...@redhat.com> wrote: On Monday, February
1, 2016,
Mykola Dvornik <mykola.dvor...@gmail.com> wrote:
Hi guys, This is sort of rebuttal. I have a CephFS deployed and
mounted on a
couple of clients via ceph-fuse (due to quota support and
possibility to
kill the ceph-fuse process to avoid stale mounts). So the problems
is that
some times the changes made on one client are not visible on the
others. It
appears to me as rather random process. The only solution is to
touch a new
file in any particular folder that apparently triggers
synchronization. I've
been using a kernel-side client before with no such kind of
problems. So the
questions is it expected behavior of ceph-fuse?
What version are you running on your servers and clients? Exactly
what
changes are you making that aren't visible? What's the output of
"ceph -s"?
We see bugs like this occasionally but I can't think of any recent
ones in
ceph-fuse -- they're actually seen a lot more often in the kernel
client.
-Greg
Regards, Mykola
_______________________________________________ ceph-users mailing
list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com