Previously we clean up the queues when we got CLOSED event. It was used to make sure we won't send leftover replies/events of a old client to a new client. Now this patch does an extra operation to flush the response queue before cleaning up.
In most cases, a QMP session will be based on a bidirectional channel (a TCP port, for example, we read/write to the same socket handle), so in port and out port of the backend chardev are fundamentally the same port. In these cases, it does not really matter much on whether we'll flush the response queue since flushing should possibly fail anyways (just imagine to write to a socket that has already closed). However there can be cases where in & out ports of the QMP monitor's backend chardev are separated. One true example: cat $QMP_COMMANDS | qemu -qmp stdio ... | filter_commands In this case, the backend is fd-typed, and it is connected to stdio where in port is stdin and out port is stdout. Now if we drop all the events on the response queue then filter_command process might miss some events that it might expect. The thing is that, when stdin closes, stdout might still be there alive! In practice, I encountered a very strange SHUTDOWN event missing when running test with iotest 087 with Out-Of-Band enabled. One condition could be this (after "quit" command is executed and QEMU quits the main loop): 1. [main thread] QEMU queues one SHUTDOWN event into response queue 2. "cat" terminates (to distinguish it from the animal, I quote it). Logically it can terminate even earlier, but let's just assume it's here. 3. [monitor iothread] QEMU reads EOF from stdin, which connects to the "cat" process 4. [monitor iothread] QEMU calls the CLOSED event hook for the monitor, which will clean up the response queue of the monitor, then the SHUTDOWN event is dropped 5. [main thread] clean up the monitors in monitor_cleanup(), when trying to flush pending responses, it sees nothing. SHUTDOWN is lost forever Note that before the monitor iothread was introduced, step [4]/[5] could never happen since the main loop was the only place to detect the EOF event of stdin and run the CLOSED event hooks. Now things can happen in parallel in the iothread. Without this patch, iotest 087 will have ~10% chance to miss the SHUTDOWN event and fail when with Out-Of-Band enabled: 087 8s ... - output mismatch (see 087.out.bad) --- /home/peterx/git/qemu/tests/qemu-iotests/087.out 2018-06-01 18:44:22.378982462 +0800 +++ /home/peterx/git/qemu/bin/tests/qemu-iotests/087.out.bad 2018-06-01 18:53:44.267840928 +0800 @@ -8,7 +8,6 @@ {"return": {}} {"error": {"class": "GenericError", "desc": "'node-name' must be specified for the root node"}} {"return": {}} -{"timestamp": {"seconds": TIMESTAMP, "microseconds": TIMESTAMP}, "event": "SHUTDOWN", "data": {"guest": false}} === Duplicate ID === @@ -53,7 +52,6 @@ {"return": {}} {"return": {}} {"return": {}} -{"timestamp": {"seconds": TIMESTAMP, "microseconds": TIMESTAMP}, "event": "SHUTDOWN", "data": {"guest": false}} This patch fixes the problem. Fixes: 6d2d563f8c ("qmp: cleanup qmp queues properly", 2018-03-27) Suggested-by: Markus Armbruster <arm...@redhat.com> Signed-off-by: Peter Xu <pet...@redhat.com> Signed-off-by: Peter Xu <pet...@redhat.com> --- monitor.c | 30 +++++++++++++++++++++++++----- 1 file changed, 25 insertions(+), 5 deletions(-) diff --git a/monitor.c b/monitor.c index 6d0cec552e..e59d4f09ac 100644 --- a/monitor.c +++ b/monitor.c @@ -512,20 +512,39 @@ struct QMPResponse { }; typedef struct QMPResponse QMPResponse; +static QObject *monitor_qmp_response_pop_one(Monitor *mon) +{ + QObject *data; + + qemu_mutex_lock(&mon->qmp.qmp_queue_lock); + data = g_queue_pop_head(mon->qmp.qmp_responses); + qemu_mutex_unlock(&mon->qmp.qmp_queue_lock); + + return data; +} + +static void monitor_qmp_response_flush(Monitor *mon) +{ + QObject *data; + + while ((data = monitor_qmp_response_pop_one(mon))) { + monitor_json_emitter_raw(mon, data); + qobject_unref(data); + } +} + /* * Return one QMPResponse. The response is only valid if * response.data is not NULL. */ -static QMPResponse monitor_qmp_response_pop_one(void) +static QMPResponse monitor_qmp_response_pop(void) { Monitor *mon; QObject *data = NULL; qemu_mutex_lock(&monitor_lock); QTAILQ_FOREACH(mon, &mon_list, entry) { - qemu_mutex_lock(&mon->qmp.qmp_queue_lock); - data = g_queue_pop_head(mon->qmp.qmp_responses); - qemu_mutex_unlock(&mon->qmp.qmp_queue_lock); + data = monitor_qmp_response_pop_one(mon); if (data) { break; } @@ -539,7 +558,7 @@ static void monitor_qmp_bh_responder(void *opaque) QMPResponse response; while (true) { - response = monitor_qmp_response_pop_one(); + response = monitor_qmp_response_pop(); if (!response.data) { break; } @@ -4366,6 +4385,7 @@ static void monitor_qmp_event(void *opaque, int event) mon_refcount++; break; case CHR_EVENT_CLOSED: + monitor_qmp_response_flush(mon); monitor_qmp_cleanup_queues(mon); json_message_parser_destroy(&mon->qmp.parser); json_message_parser_init(&mon->qmp.parser, handle_qmp_command); -- 2.17.1