Hello Cephers,

I am using CephFS, and running some benchmarks using fio.
After increasing the object_size to 33554432, when I try to run some read
and write tests with different block sizes, when I get to block size of 64m
and beyond, Ceph does not finish the operation (I tried letting it run for
more than a day at least three times).
However, when I cancel the job and I expect to see no io  operations, here
is what I get:

[cephuser@node01 ~]$ ceph -s
    cluster b7beebf6-ea9f-4560-a916-a58e106c6e8e
     health HEALTH_OK
     monmap e3: 3 mons at {node02=
192.168.17.212:6789/0,node03=192.168.17.213:6789/0,node04=192.168.17.214:6789/0
}
            election epoch 8, quorum 0,1,2 node02,node03,node04
     mdsmap e74: 1/1/1 up {0=node02=up:active}
     osdmap e324: 14 osds: 14 up, 14 in
      pgmap v155699: 768 pgs, 3 pools, 15285 MB data, 1772 objects
            91283 MB used, 7700 GB / 7817 GB avail
                 768 active+clean
*  client io 2911 MB/s rd, 90 op/s*


If I do ceph -w, it shows me that it is constantly doing reads, but I have
no idea from where and when it would stop?
I had to remove my CephFS file system and the associated pools and start
things from scratch.

1. Any idea what is happening?
2. When this happens, do you know a better way to get out of the situation
without destroying the filesystem and the pools?

Thanks,
Hadi
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to