Il 28/02/2012 11:24, Michael Tokarev ha scritto: > This removes quite some duplicated code. > > v2 fixes a bug (uninitialized reply.error) and makes the loop more natural. > > Signed-off-By: Michael Tokarev <m...@tls.msk.ru> > --- > block/nbd.c | 95 +++++++++++++++++++--------------------------------------- > 1 files changed, 31 insertions(+), 64 deletions(-) > > diff --git a/block/nbd.c b/block/nbd.c > index 161b299..a78e212 100644 > --- a/block/nbd.c > +++ b/block/nbd.c > @@ -320,91 +320,58 @@ static int nbd_open(BlockDriverState *bs, const char* > filename, int flags) > return result; > } > > -static int nbd_co_readv_1(BlockDriverState *bs, int64_t sector_num, > - int nb_sectors, QEMUIOVector *qiov, > - int offset) > -{ > - BDRVNBDState *s = bs->opaque; > - struct nbd_request request; > - struct nbd_reply reply; > - > - request.type = NBD_CMD_READ; > - request.from = sector_num * 512; > - request.len = nb_sectors * 512; > - > - nbd_coroutine_start(s, &request); > - if (nbd_co_send_request(s, &request, NULL, 0) == -1) { > - reply.error = errno; > - } else { > - nbd_co_receive_reply(s, &request, &reply, qiov->iov, offset); > - } > - nbd_coroutine_end(s, &request); > - return -reply.error; > - > -} > +/* qemu-nbd has a limit of slightly less than 1M per request. Try to > + * remain aligned to 4K. */ > +#define NBD_MAX_SECTORS 2040 > > -static int nbd_co_writev_1(BlockDriverState *bs, int64_t sector_num, > - int nb_sectors, QEMUIOVector *qiov, > - int offset) > +static int nbd_co_rwv(BlockDriverState *bs, int64_t sector_num, > + int nb_sectors, QEMUIOVector *qiov, int iswrite)
Call this nbd_co_rw, and please pass the whole request.type down. > { > BDRVNBDState *s = bs->opaque; > struct nbd_request request; > struct nbd_reply reply; > + int offset = 0; > > - request.type = NBD_CMD_WRITE; > - if (!bdrv_enable_write_cache(bs) && (s->nbdflags & NBD_FLAG_SEND_FUA)) { > + request.type = iswrite ? NBD_CMD_WRITE : NBD_CMD_READ; > + if (iswrite && !bdrv_enable_write_cache(bs) && (s->nbdflags & > NBD_FLAG_SEND_FUA)) { > request.type |= NBD_CMD_FLAG_FUA; > } > + reply.error = 0; > + > + /* we split the request into pieces of at most NBD_MAX_SECTORS size > + * and process them in a loop... */ > + do { > + request.from = sector_num * 512; > + request.len = MIN(nb_sectors, NBD_MAX_SECTORS) * 512; > + > + nbd_coroutine_start(s, &request); > + if (nbd_co_send_request(s, &request, iswrite ? qiov->iov : NULL, 0) > == -1) { The last 0 needs to be offset. > + reply.error = errno; > + } else { > + nbd_co_receive_reply(s, &request, &reply, iswrite ? NULL : > qiov->iov, offset); > + } > + nbd_coroutine_end(s, &request); > > - request.from = sector_num * 512; > - request.len = nb_sectors * 512; > + offset += NBD_MAX_SECTORS * 512; > + sector_num += NBD_MAX_SECTORS; > + nb_sectors -= NBD_MAX_SECTORS; > + > + /* ..till we hit an error or there's nothing more to process */ > + } while (reply.error == 0 && nb_sectors > 0); > > - nbd_coroutine_start(s, &request); > - if (nbd_co_send_request(s, &request, qiov->iov, offset) == -1) { > - reply.error = errno; > - } else { > - nbd_co_receive_reply(s, &request, &reply, NULL, 0); > - } > - nbd_coroutine_end(s, &request); > return -reply.error; > } > > -/* qemu-nbd has a limit of slightly less than 1M per request. Try to > - * remain aligned to 4K. */ > -#define NBD_MAX_SECTORS 2040 > - > static int nbd_co_readv(BlockDriverState *bs, int64_t sector_num, > int nb_sectors, QEMUIOVector *qiov) > { > - int offset = 0; > - int ret; > - while (nb_sectors > NBD_MAX_SECTORS) { > - ret = nbd_co_readv_1(bs, sector_num, NBD_MAX_SECTORS, qiov, offset); > - if (ret < 0) { > - return ret; > - } > - offset += NBD_MAX_SECTORS * 512; > - sector_num += NBD_MAX_SECTORS; > - nb_sectors -= NBD_MAX_SECTORS; > - } > - return nbd_co_readv_1(bs, sector_num, nb_sectors, qiov, offset); > + return nbd_co_rwv(bs, sector_num, nb_sectors, qiov, 0); > } > > static int nbd_co_writev(BlockDriverState *bs, int64_t sector_num, > int nb_sectors, QEMUIOVector *qiov) > { > - int offset = 0; > - int ret; > - while (nb_sectors > NBD_MAX_SECTORS) { > - ret = nbd_co_writev_1(bs, sector_num, NBD_MAX_SECTORS, qiov, offset); > - if (ret < 0) { > - return ret; > - } > - offset += NBD_MAX_SECTORS * 512; > - sector_num += NBD_MAX_SECTORS; > - nb_sectors -= NBD_MAX_SECTORS; > - } > - return nbd_co_writev_1(bs, sector_num, nb_sectors, qiov, offset); > + return nbd_co_rwv(bs, sector_num, nb_sectors, qiov, 1); > } ... but thinking more about it, why don't you leave nbd_co_readv_1/nbd_co_writev_1 alone, and create a nbd_split_rw function that takes a function pointer? Paolo