On Tue, Jul 10, 2018 at 2:37 AM Kevin Olbrich <k...@sv01.de> wrote: > 2018-07-10 0:35 GMT+02:00 Jason Dillaman <jdill...@redhat.com>: > >> Is the link-local address of "fe80::219:99ff:fe9e:3a86%eth0" at least >> present on the client computer you used? I would have expected the OSD to >> determine the client address, so it's odd that it was able to get a >> link-local address. >> > > Yes, it is. eth0 is part of bond0 which is a vlan trunk. Bond0.X is > attached to brX which has an ULA-prefix for the ceph cluster. > Eth0 has no address itself. In this case this must mean, the address has > been carried down to the hardware interface. > > I am wondering why it uses link local when there is an ULA-prefix > available. > > The address is available on brX on this client node. >
I'll open a tracker ticker to get that issue fixed, but in the meantime, you can run "rados -p <IMAGE POOL> rmxattr rbd_header.<IMAGE ID> lock.rbd_lock" to remove the lock. > > - Kevin > > >> On Mon, Jul 9, 2018 at 3:43 PM Kevin Olbrich <k...@sv01.de> wrote: >> >>> 2018-07-09 21:25 GMT+02:00 Jason Dillaman <jdill...@redhat.com>: >>> >>>> BTW -- are you running Ceph on a one-node computer? I thought IPv6 >>>> addresses starting w/ fe80 were link-local addresses which would probably >>>> explain why an interface scope id was appended. The current IPv6 address >>>> parser stops reading after it encounters a non hex, colon character [1]. >>>> >>> >>> No, this is a compute machine attached to the storage vlan where I >>> previously had also local disks. >>> >>> >>>> >>>> >>>> On Mon, Jul 9, 2018 at 3:14 PM Jason Dillaman <jdill...@redhat.com> >>>> wrote: >>>> >>>>> Hmm ... it looks like there is a bug w/ RBD locks and IPv6 addresses >>>>> since it is failing to parse the address as valid. Perhaps it's barfing on >>>>> the "%eth0" scope id suffix within the address. >>>>> >>>>> On Mon, Jul 9, 2018 at 2:47 PM Kevin Olbrich <k...@sv01.de> wrote: >>>>> >>>>>> Hi! >>>>>> >>>>>> I tried to convert an qcow2 file to rbd and set the wrong pool. >>>>>> Immediately I stopped the transfer but the image is stuck locked: >>>>>> >>>>>> Previusly when that happened, I was able to remove the image after 30 >>>>>> secs. >>>>>> >>>>>> [root@vm2003 images1]# rbd -p rbd_vms_hdd lock list fpi_server02 >>>>>> There is 1 exclusive lock on this image. >>>>>> Locker ID Address >>>>>> >>>>>> client.1195723 auto 93921602220416 >>>>>> [fe80::219:99ff:fe9e:3a86%eth0]:0/1200385089 >>>>>> >>>>>> [root@vm2003 images1]# rbd -p rbd_vms_hdd lock rm fpi_server02 "auto >>>>>> 93921602220416" client.1195723 >>>>>> rbd: releasing lock failed: (22) Invalid argument >>>>>> 2018-07-09 20:45:19.080543 7f6c2c267d40 -1 librados: unable to parse >>>>>> address [fe80::219:99ff:fe9e:3a86%eth0]:0/1200385089 >>>>>> 2018-07-09 20:45:19.080555 7f6c2c267d40 -1 librbd: unable to >>>>>> blacklist client: (22) Invalid argument >>>>>> >>>>>> The image is not in use anywhere! >>>>>> >>>>>> How can I force removal of all locks for this image? >>>>>> >>>>>> Kind regards, >>>>>> Kevin >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>> >>>>> >>>>> -- >>>>> Jason >>>>> >>>> >>>> [1] https://github.com/ceph/ceph/blob/master/src/msg/msg_types.cc#L108 >>>> >>>> -- >>>> Jason >>>> >>> >>> >> >> -- >> Jason >> > > -- Jason
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com