By running rbd resize <http://docs.ceph.com/docs/master/rbd/rados-rbd-cmds/> and then 'xfs_growfs -d' on the filesystem.
Is there a better way to resize an RBD image and the filesystem? On Thu, Nov 12, 2015 at 10:35 PM, Jan Schermer <j...@schermer.cz> wrote: > > On 12 Nov 2015, at 20:49, Bogdan SOLGA <bogdan.so...@gmail.com> wrote: > > Hello Jan! > > Thank you for your advices, first of all! > > The filesystem was created using mkfs.xfs, after creating the RBD block > device and mapping it on the Ceph client. I haven't specified any > parameters when I created the filesystem, I just ran mkfs.xfs on the image > name. > > As you mentioned the filesystem thinking the block device should be larger > than it is - I have initially created that image as a 2GB image, and then > resized it to be much bigger. Could this be the issue? > > > Sounds more than likely :-) How exactly did you grow it? > > Jan > > > There are several RBD images mounted on one Ceph client, but only one of > them had issues. I have made a clone, and I will try running fsck on it. > > Fortunately it's not important data, it's just testing data. If I won't > succeed repairing it I will trash and re-create it, of course. > > Thank you, once again! > > > > On Thu, Nov 12, 2015 at 9:28 PM, Jan Schermer <j...@schermer.cz> wrote: > >> How did you create filesystems and/or partitions on this RBD block device? >> The obvious causes would be >> 1) you partitioned it and the partition on which you ran mkfs points or >> pointed during mkfs outside the block device size (happens if you for >> example automate this and confuse sectors x cylinders, or if you copied the >> partition table with dd or from some image) >> or >> 2) mkfs created the filesystem with pointers outside of the block device >> for some other reason (bug?) >> or >> 3) this RBD device is a snapshot that got corrupted (or wasn't >> snapshotted in crash-consistent state and you got "lucky") and some >> reference points to a non-sensical block number (fsck could fix this, but I >> wouldn't trust the data integrity anymore) >> >> Basically the filesystem thinks the block device should be larger than it >> is and tries to reach beyond. >> >> Is this just one machine or RBD image or is there more? >> >> I'd first create a snapshot and then try running fsck on it, it should >> hopefully tell you if there's a problem in setup or a corruption. >> >> If it's not important data and it's just one instance of this problem >> then I'd just trash and recreate it. >> >> Jan >> >> On 12 Nov 2015, at 20:14, Bogdan SOLGA <bogdan.so...@gmail.com> wrote: >> >> Hello everyone! >> >> We have a recently installed Ceph cluster (v 0.94.5, Ubuntu 14.04), and >> today I noticed a lot of 'attempt to access beyond end of device' messages >> in the /var/log/syslog file. They are related to a mounted RBD image, and >> have the following format: >> >> >> *Nov 12 21:06:44 ceph-client-01 kernel: [438507.952532] attempt to access >> beyond end of deviceNov 12 21:06:44 ceph-client-01 kernel: [438507.952534] >> rbd5: rw=33, want=6193176, limit=4194304* >> >> After restarting that Ceph client, I see a lot of 'metadata I/O error' >> messages in the boot log: >> >> *XFS (rbd5): metadata I/O error: block 0x46e001 >> ("xfs_buf_iodone_callbacks") error 5 numblks 1* >> >> Any idea on why these messages are shown? The health of the cluster shows >> as OK, and I can access that block device without (apparent) issues... >> >> Thank you! >> >> Regards, >> Bogdan >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com