Thanks!

One more question, what do you mean by "bigger" ?
Do you mean that bigger block size (say, I will run read test with bs=4K,
then I need to first write the rbd with bs>4K?)? or size that is big enough
to cover the area where the test will be executed?


On Sun, Dec 11, 2016 at 9:54 PM, Somnath Roy <somnath....@sandisk.com>
wrote:

> A block needs to be written before read otherwise you will get funny
> result. For example, in case of flash (depending on how FW is implemented)
> , it will mostly return you 0 if a block is not written. Now, I have seen
> some flash FW is really inefficient on manufacturing this data (say 0) if
> not written and some are really fast.
>
> So, to get predictable result you should be always reading a block that is
> written. In a device say half of the block is written and you are doing a
> full device random reads , you will get unpredictable/spiky/imbalanced
> result.
>
> Same with rbd as well, consider it as a storage device and behavior would
> be similar. So, it is always recommended to precondition (fill up) a rbd
> image with bigger block seq write before you do any synthetic test on that.
> Now, for filestore backend added advantage of preconditioning rbd will be
> the files in the filesystem will be created beforehand.
>
>
>
> Thanks & Regards
>
> Somnath
>
>
>
> *From:* V Plus [mailto:v.plussh...@gmail.com]
> *Sent:* Sunday, December 11, 2016 6:01 PM
> *To:* Somnath Roy
> *Cc:* ceph-users@lists.ceph.com
> *Subject:* Re: [ceph-users] Ceph performance is too good (impossible..)...
>
>
>
> Thanks Somnath!
>
> As you recommended, I executed:
>
> dd if=/dev/zero bs=1M count=4096 of=/dev/rbd0
>
> dd if=/dev/zero bs=1M count=4096 of=/dev/rbd1
>
>
>
> Then the output results look more reasonable!
>
> Could you tell me why??
>
>
>
> Btw, the purpose of my run is to test the performance of rbd in ceph. Does
> my case mean that before every test, I have to "initialize" all the
> images???
>
>
>
> Great thanks!!
>
>
>
> On Sun, Dec 11, 2016 at 8:47 PM, Somnath Roy <somnath....@sandisk.com>
> wrote:
>
> Fill up the image with big write (say 1M) first before reading and you
> should see sane throughput.
>
>
>
> Thanks & Regards
>
> Somnath
>
> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf
> Of *V Plus
> *Sent:* Sunday, December 11, 2016 5:44 PM
> *To:* ceph-users@lists.ceph.com
> *Subject:* [ceph-users] Ceph performance is too good (impossible..)...
>
>
>
> Hi Guys,
>
> we have a ceph cluster with 6 machines (6 OSD per host).
>
> 1. I created 2 images in Ceph, and map them to another host A (*outside *the
> Ceph cluster). On host A, I got */dev/rbd0* and* /dev/rbd1*.
>
> 2. I start two fio job to perform READ test on rbd0 and rbd1. (fio job
> descriptions can be found below)
>
> *"sudo fio fioA.job -output a.txt & sudo  fio fioB.job -output b.txt  &
> wait"*
>
> 3. After the test, in a.txt, we got *bw=1162.7MB/s*, in b.txt, we get
> *bw=3579.6MB/s*.
>
> The results do NOT make sense because there is only one NIC on host A, and
> its limit is 10 Gbps (1.25GB/s).
>
>
>
> I suspect it is because of the cache setting.
>
> But I am sure that in file */etc/ceph/ceph.conf* on host A,I already
> added:
>
> *[client]*
>
> *rbd cache = false*
>
>
>
> Could anyone give me a hint what is missing? why....
>
> Thank you very much.
>
>
>
> *fioA.job:*
>
> *[A]*
>
> *direct=1*
>
> *group_reporting=1*
>
> *unified_rw_reporting=1*
>
> *size=100%*
>
> *time_based=1*
>
> *filename=/dev/rbd0*
>
> *rw=read*
>
> *bs=4MB*
>
> *numjobs=16*
>
> *ramp_time=10*
>
> *runtime=20*
>
>
>
> *fioB.job:*
>
> *[B]*
>
> *direct=1*
>
> *group_reporting=1*
>
> *unified_rw_reporting=1*
>
> *size=100%*
>
> *time_based=1*
>
> *filename=/dev/rbd1*
>
> *rw=read*
>
> *bs=4MB*
>
> *numjobs=16*
>
> *ramp_time=10*
>
> *runtime=20*
>
>
>
> *Thanks...*
>
> PLEASE NOTE: The information contained in this electronic mail message is
> intended only for the use of the designated recipient(s) named above. If
> the reader of this message is not the intended recipient, you are hereby
> notified that you have received this message in error and that any review,
> dissemination, distribution, or copying of this message is strictly
> prohibited. If you have received this communication in error, please notify
> the sender by telephone or e-mail (as shown above) immediately and destroy
> any and all copies of this message in your possession (whether hard copies
> or electronically stored copies).
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to