I am seeing the same thing, and was wondering the same. We have 16 OSDs on 4 hosts. The File system is Xfs. The OS is CentOS 6.4. ceph version 0.72.2
I am importing a 3.3TB disk image into a rbd image. At 2.6TB, and still importing, 5.197TB is used according to `rados -p <pool> df` With previous images I have imported, they were small and imported so quickly I did not get to check disk utilization during import. After import however, `rados -p <pool> df` reported utilization (the difference before and after import) about the same as the image size that was imported. -RG ----- Original Message ----- From: "Mark Kirkwood" <mark.kirkw...@catalyst.net.nz> To: "Gregory Farnum" <g...@inktank.com> Cc: ceph-users@lists.ceph.com Sent: Thursday, April 10, 2014 5:36:11 AM Subject: Re: [ceph-users] OSD space usage 2x object size after rados put Some more checking: - re-deploying the cluster and testing again - same result (initial 2x space usage). - re-deploying with ext4 for OSD's (instead of default xfs)...*no* 2x space usage observed. Retested several times. So looks like some combination of xfs/kernel/os version (Ubuntu 13.10) weirdness. Regards Mark On 10/04/14 15:41, Mark Kirkwood wrote: > Redoing (attached, 1st file is for 2x space, 2nd for normal). I'm seeing: > > $ diff osd-du.0.txt osd-du.1.txt > 924,925c924,925 > < 2048 /var/lib/ceph/osd/ceph-1/current/5.1a_head/file__head_2E6FB49A__5 > < 2048 /var/lib/ceph/osd/ceph-1/current/5.1a_head > --- > > 1024 /var/lib/ceph/osd/ceph-1/current/5.1a_head/file__head_2E6FB49A__5 > > 1024 /var/lib/ceph/osd/ceph-1/current/5.1a_head > 931c931 > < 2054 /var/lib/ceph/osd/ceph-1/current > --- > > 1030 /var/lib/ceph/osd/ceph-1/current > 936c936 > < 2054 /var/lib/ceph/osd/ceph-1/ > --- > > 1030 /var/lib/ceph/osd/ceph-1/ > > Looks like the actual object has twice the disk footprint. > Interestingly, comparing du vs ls info for it at that point shows: > > $ ls -l > total 2097088 > -rw-r--r-- 1 root root 1073741824 Apr 10 15:33 file__head_2E6FB49A__5 > > $ du file__head_2E6FB49A__5 > 2097088 file__head_2E6FB49A__5 > > ...which is interesting. > > Regards > > Mark > > On 10/04/14 15:11, Mark Kirkwood wrote: >> Ah right - sorry, I didn't realize that my 'du' was missing the files! >> I will retest and post updated output. >> >> Cheers >> >> Mark >> >> On 10/04/14 15:04, Gregory Farnum wrote: >>> Right, but I'm interested in the space allocation within the PG. The >>> best guess I can come up with without trawling through the code is >>> that some layer in the stack is preallocated and then trimmed the >>> objects back down once writing stops, but I'd like some more data >>> points before I dig. >>> -Greg >>> Software Engineer #42 @ http://inktank.com | http://ceph.com >>> >>> >>> On Wed, Apr 9, 2014 at 7:59 PM, Mark Kirkwood >>> <mark.kirkw...@catalyst.net.nz> wrote: >>>> It is only that single pg using the space (see attached) - but >>>> essentially: >>>> >>>> $ du -m /var/lib/ceph/osd/ceph-1 >>>> ... >>>> 2048 /var/lib/ceph/osd/ceph-1/current/5.1a_head >>>> 2053 /var/lib/ceph/osd/ceph-1/current >>>> 2053 /var/lib/ceph/osd/ceph-1/ >>>> >>>> Which is resized to 1025 soon after. Interestingly I am not seeing this >>>> effect (same ceph version) on a single host setup with 2 osds using >>>> preexisting partitions... it's only on these multi host >>>> configurations that >>>> have osd's using whole devices (both setups installed using >>>> ceph-deploy, so >>>> in theory nothing exotic about 'em except for the multi 'hosts' are >>>> actually >>>> VMs). >>>> >>>> Regards >>>> >>>> Mark >>>> >>>> On 10/04/14 02:27, Gregory Farnum wrote: >>>>> I don't think the backing store should be seeing any effects like >>>>> that. What are the filenames which are using up that space inside the >>>>> folders? >>>>> -Greg >>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com >>>>> >>>>> >>>>> On Wed, Apr 9, 2014 at 1:58 AM, Mark Kirkwood >>>>> <mark.kirkw...@catalyst.net.nz> wrote: >>>>>> Hi all, >>>>>> >>>>>> I've noticed that objects are using twice their actual space for a >>>>>> few >>>>>> minutes after they are 'put' via rados: >>>>>> >>>>>> $ ceph -v >>>>>> ceph version 0.79-42-g010dff1 >>>>>> (010dff12c38882238591bb042f8e497a1f7ba020) >>>>>> >>>>>> $ ceph osd tree >>>>>> # id weight type name up/down reweight >>>>>> -1 0.03998 root default >>>>>> -2 0.009995 host ceph1 >>>>>> 0 0.009995 osd.0 up 1 >>>>>> -3 0.009995 host ceph2 >>>>>> 1 0.009995 osd.1 up 1 >>>>>> -4 0.009995 host ceph3 >>>>>> 2 0.009995 osd.2 up 1 >>>>>> -5 0.009995 host ceph4 >>>>>> 3 0.009995 osd.3 up 1 >>>>>> >>>>>> $ ceph osd dump|grep repool >>>>>> pool 5 'repool' replicated size 3 min_size 2 crush_ruleset 0 >>>>>> object_hash >>>>>> rjenkins pg_num 64 pgp_num 64 last_change 57 owner 0 flags hashpspool >>>>>> stripe_width 0 >>>>>> >>>>>> $ du -m file >>>>>> 1025 file >>>>>> >>>>>> $ rados put -p repool file file >>>>>> >>>>>> $ cd /var/lib/ceph/osd/ceph-1/current/ >>>>>> $ du -m 5.1a_head >>>>>> 2048 5.1a_head >>>>>> >>>>>> [later] >>>>>> >>>>>> $ du -m 5.1a_head >>>>>> 1024 5.1a_head >>>>>> >>>>>> The above situation is repeated on the other two OSD's where this >>>>>> pg is >>>>>> mapped. So after about 5 minutes or so we have (as expected) that >>>>>> the 1G >>>>>> file is using 1G on each of the 3 OSD's it is mapped to, however >>>>>> for a >>>>>> short >>>>>> period of time it is using twice this! I very interested to know what >>>>>> activity is happening that causes the 2x space use - as this could >>>>>> be a >>>>>> significant foot gun if uploading large files when we don't have >>>>>> 2x the >>>>>> space available on each OSD. >>>>>> >>>>>> Regards >>>>>> >>>>>> Mark >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com