On 24.07.23 15:57, Andrey Zhadchenko wrote:
On 7/24/23 14:20, Alexander Atanasov wrote:
On 24.07.23 11:03, Konstantin Khorenko wrote:
Implement find_hole() for dm-qcow2 target.
Iterate over ranges with cluster granularity until hole or data is
found.
To reduce code duplication, we should use already existing
parse_metadata()
We can pretend that seek request is read request for metadata purposes
and than interpret parsing result in our favor.
Since parse_metadata() support request postponing (for example when the
requested L2 cluster is absent in RAM), we should create separate qio
list for our queries.
https://jira.vzint.dev/browse/PSBM-145746
Signed-off-by: Andrey Zhadchenko <andrey.zhadche...@virtuozzo.com>
---
drivers/md/dm-qcow2-map.c | 140 +++++++++++++++++++++++++++++++++++
drivers/md/dm-qcow2-target.c | 1 +
drivers/md/dm-qcow2.h | 2 +
3 files changed, 143 insertions(+)
diff --git a/drivers/md/dm-qcow2-map.c b/drivers/md/dm-qcow2-map.c
index a779889c6970..f728a52ab5e4 100644
--- a/drivers/md/dm-qcow2-map.c
+++ b/drivers/md/dm-qcow2-map.c
@@ -3980,6 +3980,14 @@ static void process_resubmit_qios(struct qcow2
*qcow2, struct list_head *qios)
}
}
+static void process_seek_qios(struct qcow2 *qcow, struct list_head
*qios)
+{
+ struct qio *qio;
+
+ while ((qio = qio_list_pop(qios)) != NULL)
+ complete(qio->data);
+}
+
void do_qcow2_work(struct work_struct *ws)
{
struct qcow2 *qcow2 = container_of(ws, struct qcow2, worker);
@@ -3991,6 +3999,7 @@ void do_qcow2_work(struct work_struct *ws)
LIST_HEAD(cow_indexes_qios);
LIST_HEAD(cow_end_qios);
LIST_HEAD(resubmit_qios);
+ LIST_HEAD(seek_qios);
unsigned int pflags = current->flags;
current->flags |= PF_LOCAL_THROTTLE|PF_MEMALLOC_NOIO;
@@ -4003,6 +4012,7 @@ void do_qcow2_work(struct work_struct *ws)
list_splice_init(&qcow2->qios[QLIST_COW_INDEXES],
&cow_indexes_qios);
list_splice_init(&qcow2->qios[QLIST_COW_END], &cow_end_qios);
list_splice_init(&qcow2->resubmit_qios, &resubmit_qios);
+ list_splice_init(&qcow2->qios[QLIST_SEEK], &seek_qios);
spin_unlock_irq(&qcow2->deferred_lock);
process_embedded_qios(qcow2, &embedded_qios, &deferred_qios);
@@ -4013,6 +4023,7 @@ void do_qcow2_work(struct work_struct *ws)
process_cow_indexes_write(qcow2, &cow_indexes_qios);
process_cow_end(qcow2, &cow_end_qios);
process_resubmit_qios(qcow2, &resubmit_qios);
+ process_seek_qios(qcow2, &seek_qios);
/* This actually submits batch of md writeback, initiated
above */
submit_metadata_writeback(qcow2);
@@ -4235,3 +4246,132 @@ static void handle_cleanup_mask(struct qio *qio)
ext->cleanup_mask &= ~FREE_ALLOCATED_CLU;
}
}
+
+static sector_t get_next_l2(struct qio *qio)
+{
+ struct qcow2 *qcow2 = qio->qcow2;
+ loff_t start, add;
+
+ start = to_bytes(qio->bi_iter.bi_sector);
+ add = qcow2->l2_entries - (start / qcow2->clu_size) %
qcow2->l2_entries;
+
+ return qio->bi_iter.bi_sector + (qcow2->clu_size / to_bytes(1))
* add;
+}
+
+static sector_t get_next_clu(struct qio *qio)
+{
+ struct qcow2 *qcow2 = qio->qcow2;
+ loff_t offset;
+
+ offset = to_bytes(qio->bi_iter.bi_sector);
+ offset = offset / qcow2->clu_size;
+ offset = (offset + 1) * qcow2->clu_size;
+
Isn't
offset = (offset + qcow2->clu_size) / qcow2->clu_size;
offset *= qcow2->clu_size;
more optimal ? If clu_size is known to be power of two it may be
possible to rewritte it in another way.
I think it is the same three operations of sum, division and multiplication
As far as I am aware, cluster size is not guaranteed to be a power of two
Academically you are correct but there is still a small difference - one
register load.
+ return to_sector(offset);
+}
+
+loff_t qcow2_find_hole(struct dm_target *ti, loff_t offset, int whence)
+{
+ struct qcow2 *qcow2 = to_qcow2_target(ti)->top;
+ DECLARE_COMPLETION_ONSTACK(compl);
+ bool unmapped, zeroes, try_lower;
+ struct qio qio = {0}, *qptr;
+ loff_t result = -EINVAL;
+ struct qcow2_map map;
+ u32 size;
+ int ret;
+
+ qio.bi_iter.bi_sector = to_sector(offset);
+ qio.bi_iter.bi_size = qcow2->clu_size - offset % qcow2->clu_size;
+
+ qcow2_init_qio(&qio, REQ_OP_READ, qcow2);
+ qio.queue_list_id = QLIST_SEEK;
+ qio.data = &compl;
+
+ while (qio.bi_iter.bi_sector < to_sector(qcow2->hdr.size)) {
+ qio.qcow2 = qcow2;
+retry:
+ memset(&map, 0, sizeof(map));
+ map.qcow2 = qio.qcow2;
+ qptr = &qio;
+ ret = parse_metadata(qio.qcow2, &qptr, &map);
+ /* ENXIO has a special meaning for llseek so remap it to
EINVAL*/
+ if (ret < 0)
+ return (ret == -ENXIO) ? -EINVAL : ret;
+ if (qptr == NULL) {
+ wait_for_completion(&compl);
+ reinit_completion(&compl);
What's the point of this ? compl is local , it is assigned to qio.data,
who will complete it here after qptr is already null and
parse_metadata is done? Looks like the places that use completion
manage their own completion function and data - submit_rw_md_page,
perform_rw_mapped, submit_read_whole_cow_clu.
parse_metadata() -> __handle_md_page() -> qcow2_md_page_find_or_postpone()
here it is checked if L1 or L2 table (which is also a cluster) is loaded
from disk. If it is not, the qio is inserted into md->wait_list, which
is popped after the mapping is loaded. Qio goes into qio.queue_list_id
queue. After that process_seek_qios() is called which releases completion.
So parse_metadata() automatically postpone the request until the
metadata pages are loaded and we wait on completion for it. After that
completion is rearmed for further requests.
yes, got it - i've missed that a new qio is allocated and that the
initial one goes into the wait list.
+ goto retry;
+ }
+
+calc_subclu:
+ zeroes = unmapped = try_lower = false;
+ zeroes = (size = qio_all_zeroes_size(qio.qcow2, &qio, &map));
+ if (!size)
+ unmapped = (size = qio_unmapped_size(qio.qcow2, &qio,
&map));
+ if (!size)
+ size = qio_mapped_not_zeroes_size(qio.qcow2, &qio, &map);
Why two if (!size) ?
+ if (unmapped)
+ try_lower = maybe_mapped_in_lower_delta(qio.qcow2, &qio);
This above needs comments more comments - why it is done that way ?
+
+ if (unmapped && try_lower) {
+ loff_t end = to_bytes(qio.bi_iter.bi_sector) +
qio.bi_iter.bi_size;
+
+ if (end < qio.qcow2->hdr.size) {
+ qio.qcow2 = qio.qcow2->lower;
+ goto retry;
+ }
+ }
+
+ if (whence & SEEK_HOLE) {
+ if (zeroes || unmapped) {
+ result = to_bytes(qio.bi_iter.bi_sector);
+ break;
+ } else if (size != qio.bi_iter.bi_size) {
+ /*
+ * range starts with data subclusters and after that
+ * some subclusters are zero or unmapped
+ */
+ result = to_bytes(qio.bi_iter.bi_sector) + size;
+ break;
+ }
+ }
+
+ if (whence & SEEK_DATA) {
+ if (!zeroes && !unmapped) {
+ result = to_bytes(qio.bi_iter.bi_sector);
+ break;
+ } else if (size != qio.bi_iter.bi_size) {
+ /*
+ * range starts with zero or unmapped subclusters
+ * but after that it still can be unmapped or zero
+ * We do not need to parse metadata again but we should
+ * skip this sublusters and look onto next ones
+ */
+ qio.bi_iter.bi_sector += to_sector(size);
+ qio.bi_iter.bi_size -= size;
+ goto calc_subclu;
+ }
+ }
+
+ /* whole L2 table is unmapped - skip to next l2 table */
+ if (!(map.level & L2_LEVEL))
+ qio.bi_iter.bi_sector = get_next_l2(&qio);
+ else
+ qio.bi_iter.bi_sector = get_next_clu(&qio);
+
+ qio.bi_iter.bi_size = qcow2->clu_size;
+ }
+
This whole retry logic is very long and it is hard to follow, could it
be split into functions ?
Probably, but it will look miserably due to bunch of 2-3 line functions
with 5+ arguments and introduce recursion (which will add extra pain
a helper struct instead of the 5 arguments can do.
with different level qios and its completions), as for some mappings we
need to go deeper into qcow2 backing images.
Also subclusters add a lot of hassle unfortunately.
Inspiration was taken from process_read_qio().
Also, maybe you should pinpoint some hard-to-follow places so I can add
more explaining comments?
Probably first goto retry after completion can be replaced by continue
and retry renamed into try_lower_image, if it will ease your pain :)
I am okay with it. But i am worried about you looking at that code after
an year :)
+ if (result >= 0 && result < offset)
+ result = offset;
+
+ if (qio.bi_iter.bi_sector >= to_sector(qcow2->hdr.size)) {
+ if (whence & SEEK_HOLE)
+ result = qcow2->hdr.size;
+ if (whence & SEEK_DATA)
+ result = -ENXIO;
+ }
+
+ return result;
+}
Ok, with the explaination of wait_list - LGTM.
--
Regards,
Alexander Atanasov
_______________________________________________
Devel mailing list
Devel@openvz.org
https://lists.openvz.org/mailman/listinfo/devel