Can you post the output of:

blockdev --getsz --getss --getbsz /dev/rbd5
and
xfs_info /dev/rbd5

rbd resize can actually (?) shrink the image as well - is it possible that the 
device was actually larger and you shrunk it?

Jan

> On 12 Nov 2015, at 21:46, Bogdan SOLGA <bogdan.so...@gmail.com> wrote:
> 
> By running rbd resize <http://docs.ceph.com/docs/master/rbd/rados-rbd-cmds/> 
> and then 'xfs_growfs -d' on the filesystem.
> 
> Is there a better way to resize an RBD image and the filesystem?
> 
> On Thu, Nov 12, 2015 at 10:35 PM, Jan Schermer <j...@schermer.cz 
> <mailto:j...@schermer.cz>> wrote:
> 
>> On 12 Nov 2015, at 20:49, Bogdan SOLGA <bogdan.so...@gmail.com 
>> <mailto:bogdan.so...@gmail.com>> wrote:
>> 
>> Hello Jan!
>> 
>> Thank you for your advices, first of all!
>> 
>> The filesystem was created using mkfs.xfs, after creating the RBD block 
>> device and mapping it on the Ceph client. I haven't specified any parameters 
>> when I created the filesystem, I just ran mkfs.xfs on the image name.
>> 
>> As you mentioned the filesystem thinking the block device should be larger 
>> than it is - I have initially created that image as a 2GB image, and then 
>> resized it to be much bigger. Could this be the issue?
> 
> Sounds more than likely :-) How exactly did you grow it?
> 
> Jan
> 
>> 
>> There are several RBD images mounted on one Ceph client, but only one of 
>> them had issues. I have made a clone, and I will try running fsck on it.
>> 
>> Fortunately it's not important data, it's just testing data. If I won't 
>> succeed repairing it I will trash and re-create it, of course.
>> 
>> Thank you, once again!
>> 
>> 
>> 
>> On Thu, Nov 12, 2015 at 9:28 PM, Jan Schermer <j...@schermer.cz 
>> <mailto:j...@schermer.cz>> wrote:
>> How did you create filesystems and/or partitions on this RBD block device?
>> The obvious causes would be
>> 1) you partitioned it and the partition on which you ran mkfs points or 
>> pointed during mkfs outside the block device size (happens if you for 
>> example automate this and confuse sectors x cylinders, or if you copied the 
>> partition table with dd or from some image)
>> or
>> 2) mkfs created the filesystem with pointers outside of the block device for 
>> some other reason (bug?)
>> or
>> 3) this RBD device is a snapshot that got corrupted (or wasn't snapshotted 
>> in crash-consistent state and you got "lucky") and some reference points to 
>> a non-sensical block number (fsck could fix this, but I wouldn't trust the 
>> data integrity anymore)
>> 
>> Basically the filesystem thinks the block device should be larger than it is 
>> and tries to reach beyond.
>> 
>> Is this just one machine or RBD image or is there more?
>> 
>> I'd first create a snapshot and then try running fsck on it, it should 
>> hopefully tell you if there's a problem in setup or a corruption.
>> 
>> If it's not important data and it's just one instance of this problem then 
>> I'd just trash and recreate it.
>> 
>> Jan
>> 
>>> On 12 Nov 2015, at 20:14, Bogdan SOLGA <bogdan.so...@gmail.com 
>>> <mailto:bogdan.so...@gmail.com>> wrote:
>>> 
>>> Hello everyone!
>>> 
>>> We have a recently installed Ceph cluster (v 0.94.5, Ubuntu 14.04), and 
>>> today I noticed a lot of 'attempt to access beyond end of device' messages 
>>> in the /var/log/syslog file. They are related to a mounted RBD image, and 
>>> have the following format:
>>> 
>>> Nov 12 21:06:44 ceph-client-01 kernel: [438507.952532] attempt to access 
>>> beyond end of device
>>> Nov 12 21:06:44 ceph-client-01 kernel: [438507.952534] rbd5: rw=33, 
>>> want=6193176, limit=4194304
>>> 
>>> After restarting that Ceph client, I see a lot of 'metadata I/O error' 
>>> messages in the boot log:
>>> 
>>> XFS (rbd5): metadata I/O error: block 0x46e001 ("xfs_buf_iodone_callbacks") 
>>> error 5 numblks 1
>>> 
>>> Any idea on why these messages are shown? The health of the cluster shows 
>>> as OK, and I can access that block device without (apparent) issues...
>>> 
>>> Thank you!
>>> 
>>> Regards,
>>> Bogdan
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
>>> <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>> 
>> 
> 
> 

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to