We now send several pages at a time each time that we wakeup a thread. Signed-off-by: Juan Quintela <quint...@redhat.com>
-- Use iovec's instead of creating the equivalent. Clear memory used by pages (dave) Use g_new0(danp) define MULTIFD_CONTINUE --- migration/ram.c | 57 ++++++++++++++++++++++++++++++++++++++++++++++++--------- 1 file changed, 48 insertions(+), 9 deletions(-) diff --git a/migration/ram.c b/migration/ram.c index 03f3427..7310da9 100644 --- a/migration/ram.c +++ b/migration/ram.c @@ -49,6 +49,7 @@ #include "migration/colo.h" #include "sysemu/sysemu.h" #include "qemu/uuid.h" +#include "qemu/iov.h" /***********************************************************/ /* ram save/restore */ @@ -362,6 +363,15 @@ static void compress_threads_save_setup(void) /* Multiple fd's */ +/* used to continue on the same multifd group */ +#define MULTIFD_CONTINUE UINT16_MAX + +typedef struct { + int num; + size_t size; + struct iovec *iov; +} multifd_pages_t; + struct MultiFDSendParams { /* not changed */ uint8_t id; @@ -371,11 +381,7 @@ struct MultiFDSendParams { QemuMutex mutex; /* protected by param mutex */ bool quit; - /* This is a temp field. We are using it now to transmit - something the address of the page. Later in the series, we - change it for the real page. - */ - uint8_t *address; + multifd_pages_t pages; /* protected by multifd mutex */ /* has the thread finish the last submitted job */ bool done; @@ -388,8 +394,24 @@ struct { int count; QemuMutex mutex; QemuSemaphore sem; + multifd_pages_t pages; } *multifd_send_state; +static void multifd_init_group(multifd_pages_t *pages) +{ + pages->num = 0; + pages->size = migrate_multifd_group(); + pages->iov = g_new0(struct iovec, pages->size); +} + +static void multifd_clear_group(multifd_pages_t *pages) +{ + pages->num = 0; + pages->size = 0; + g_free(pages->iov); + pages->iov = NULL; +} + static void terminate_multifd_send_threads(void) { int i; @@ -419,9 +441,11 @@ void multifd_save_cleanup(void) qemu_mutex_destroy(&p->mutex); qemu_sem_destroy(&p->sem); socket_send_channel_destroy(p->c); + multifd_clear_group(&p->pages); } g_free(multifd_send_state->params); multifd_send_state->params = NULL; + multifd_clear_group(&multifd_send_state->pages); g_free(multifd_send_state); multifd_send_state = NULL; } @@ -454,8 +478,8 @@ static void *multifd_send_thread(void *opaque) qemu_mutex_unlock(&p->mutex); break; } - if (p->address) { - p->address = 0; + if (p->pages.num) { + p->pages.num = 0; qemu_mutex_unlock(&p->mutex); qemu_mutex_lock(&multifd_send_state->mutex); p->done = true; @@ -484,6 +508,7 @@ int multifd_save_setup(void) multifd_send_state->count = 0; qemu_mutex_init(&multifd_send_state->mutex); qemu_sem_init(&multifd_send_state->sem, 0); + multifd_init_group(&multifd_send_state->pages); for (i = 0; i < thread_count; i++) { char thread_name[16]; MultiFDSendParams *p = &multifd_send_state->params[i]; @@ -493,7 +518,7 @@ int multifd_save_setup(void) p->quit = false; p->id = i; p->done = true; - p->address = 0; + multifd_init_group(&p->pages); p->c = socket_send_channel_create(); if (!p->c) { error_report("Error creating a send channel"); @@ -512,6 +537,17 @@ static uint16_t multifd_send_page(uint8_t *address, bool last_page) { int i; MultiFDSendParams *p = NULL; /* make happy gcc */ + multifd_pages_t *pages = &multifd_send_state->pages; + + pages->iov[pages->num].iov_base = address; + pages->iov[pages->num].iov_len = TARGET_PAGE_SIZE; + pages->num++; + + if (!last_page) { + if (pages->num < (pages->size - 1)) { + return MULTIFD_CONTINUE; + } + } qemu_sem_wait(&multifd_send_state->sem); qemu_mutex_lock(&multifd_send_state->mutex); @@ -525,7 +561,10 @@ static uint16_t multifd_send_page(uint8_t *address, bool last_page) } qemu_mutex_unlock(&multifd_send_state->mutex); qemu_mutex_lock(&p->mutex); - p->address = address; + p->pages.num = pages->num; + iov_copy(p->pages.iov, pages->num, pages->iov, pages->num, 0, + iov_size(pages->iov, pages->num)); + pages->num = 0; qemu_mutex_unlock(&p->mutex); qemu_sem_post(&p->sem); -- 2.9.4