>[CC-ing qemu-block, Vladimir, Kevin, and John – when sending patches, >please look into the MAINTAINERS file or use the >scripts/get_maintainer.pl script to find out who to CC on them. It’s >very to overlook patches on qemu-devel :/] > >On 07.12.21 11:56, Yi Wang wrote: >> From: Long YunJian <long.yunj...@zte.com.cn> >> >> when blockcommit from active leaf node, sometimes, we get assertion failed >> with >> "mirror_run: Assertion `QLIST_EMPTY(&bs->tracked_requests)' failed" messages. >> According to the core file, we find bs->tracked_requests has IO request, >> so assertion failed. >> (gdb) bt >> #0 0x00007f410df707cf in raise () from /lib64/libc.so.6 >> #1 0x00007f410df5ac05 in abort () from /lib64/libc.so.6 >> #2 0x00007f410df5aad9 in __assert_fail_base.cold.0 () from /lib64/libc..so.6 >> #3 0x00007f410df68db6 in __assert_fail () from /lib64/libc.so.6 >> #4 0x0000556915635371 in mirror_run (job=0x556916ff8600, errp=<optimized >> out>) at block/mirror.c:1092 >> #5 0x00005569155e6c53 in job_co_entry (opaque=0x556916ff8600) at job..c:904 >> #6 0x00005569156d9483 in coroutine_trampoline (i0=<optimized out>, >> i1=<optimized out>) at util/coroutine-ucontext.c:115 >> (gdb) p s->mirror_top_bs->backing->bs->tracked_requests >> $12 = {lh_first = 0x7f3f07bfb8b0} >> (gdb) p s->mirror_top_bs->backing->bs->tracked_requests->lh_first >> $13 = (struct BdrvTrackedRequest *) 0x7f3f07bfb8b0 >> >> Actually, before excuting assert(QLIST_EMPTY(&bs->tracked_requests)), >> it will excute mirror_flush(s). It may handle new I/O request and maybe >> pending I/O during this flush. Just likes in bdrv_close fuction, >> bdrv_drain(bs) followed by bdrv_flush(bs), we should add bdrv_drain fuction >> to handle pending I/O after mirror_flush. > >Oh. How is that happening, though? I would have expected that flushing >the target BB (and associated BDS) only flushes requests to the OS and >lower layers, but the source node (which is `bs`) should (in the case of >commit) always be above the target, so I wouldn’t have expected it to >get any new requests due to this flush. > >Do you have a reproducer for this?
As i know, flush maybe will do some thring write, and then in qcow2_co_pwritev function, if others aready hold "s->lock" lock, qemu_co_mutex_lock(&s->lock) will go to qemu_coroutine_yield, and do some other things. Maybe, it will handle new I/O now. reproducer: first, create leaf alpine-min_leaf1.qcow2, alpine-min_leaf1.qcow2 has backing file alpine-min.qcow2, and create vm from leaf qcow2, xml likes this <disk type='file' device='disk'> <driver name='qemu' type='qcow2' cache='none' iothread='1'/> <source file='/home/alpine-min_leaf1.qcow2'/> <target dev='vda' bus='virtio'/> </disk> and then make I/O request by read and write files all the time in guset os. finally, run script blockcommit.sh until assertion failed. You should modify script according to actual situation. script as follows: #!/bin/bash QEMUIMG=/home/qemu-6.0.0-rc2/build/qemu-img VIRSH=/home/libvirt-7.2.0/build/tools/virsh DOMID=v6_host7 LEAF1=/home/alpine-min_leaf1.qcow2 LEAF2=/home/alpine-min_leaf2.qcow2 while [ 1 ] do [ -f quitfile ] && break ${QEMUIMG} create -f qcow2 -F qcow2 -b ${LEAF1} ${LEAF2} sleep 2 ${VIRSH} -k 0 snapshot-create-as ${DOMID} --disk-only --diskspec vda,file=${LEAF2} --reuse-external --no-metadata [ $? -ne 0 ] && break sleep 30 ${VIRSH} blockcommit --domain ${DOMID} vda --base ${LEAF1} --top ${LEAF2} --wait --verbose --active --pivot [ $? -ne 0 ] && break done The probability of the problem is not high, it's difficult to reproduce the problem, blockcommit hundreds of times at least. >> Signed-off-by: Long YunJian <long.yunj...@zte.com.cn> >> Signed-off-by: Yi Wang <wang.y...@zte.com.cn> >> --- >> block/mirror.c | 2 ++ >> 1 file changed, 2 insertions(+) >> >> diff --git a/block/mirror.c b/block/mirror.c >> index efec2c7674..1eec356310 100644 >> --- a/block/mirror.c >> +++ b/block/mirror.c >> @@ -1079,6 +1079,8 @@ static int coroutine_fn mirror_run(Job *job, Error >> **errp) >> s->in_drain = false; >> continue; >> } >> + /* in case flush left pending I/O */ >> + bdrv_drain(bs); > >I don’t think this works, because if we drain, we would also need to >flush the target again. Essentially I believe we’d basically need >something like > >do { > bdrv_drained_begin(bs); > mirror_flush(s); > if (!QLIST_EMPTY(&bs->tracked_requests)) { > bdrv_drained_end(bs); > } >} while (!QLIST_EMPTY(&bs->tracked_requests)); > >(Which I know is really ugly) > >Hanna Yes, add bdrv_drain(bs) after mirror_flush(s) can't fix this problem. Thank you for your idea of a solution. and I'd like to modify as follows, how do you like it? @@ -1074,7 +1074,7 @@ static int coroutine_fn mirror_run(Job *job, Error **errp) s->in_drain = true; bdrv_drained_begin(bs); cnt = bdrv_get_dirty_count(s->dirty_bitmap); - if (cnt > 0 || mirror_flush(s) < 0) { + if (cnt > 0 || mirror_flush(s) < 0 || !QLIST_EMPTY(&bs->tracked_requests)) { bdrv_drained_end(bs); s->in_drain = false; continue; > > /* The two disks are in sync. Exit and report successful > * completion.