Am 22.02.2014 um 14:00 hat Peter Lieven geschrieben:
this patch tries to optimize zero write requests
by automatically using bdrv_write_zeroes if it is
supported by the format.
i know that there is a lot of potential for discussion, but i would
like to know what the others think.
this should significantly speed up file system initialization and
should speed zero write test used to test backend storage performance.
the difference can simply be tested by e.g.
dd if=/dev/zero of=/dev/vdX bs=1M
Signed-off-by: Peter Lieven <address@hidden>
As you probably have expected, there's no way I can let the patch in in
this form. The least you need to introduce is a boolean option to enable
or disable the zero check. (The default would probably be disabled, but
we can discuss this.)
block.c | 8 ++++++++
include/qemu-common.h | 1 +
util/iov.c | 20 ++++++++++++++++++++
3 files changed, 29 insertions(+)
diff --git a/block.c b/block.c
index 6f4baca..505888e 100644
--- a/block.c
+++ b/block.c
@@ -3145,6 +3145,14 @@ static int coroutine_fn
bdrv_aligned_pwritev(BlockDriverState *bs,
ret = notifier_with_return_list_notify(&bs->before_write_notifiers, req);
+ if (!ret && !(flags & BDRV_REQ_ZERO_WRITE) &&
+ drv->bdrv_co_write_zeroes && qemu_iovec_is_zero(qiov)) {
+ flags |= BDRV_REQ_ZERO_WRITE;
+ /* if the device was not opened with discard=on the below flag
+ * is immediately cleared again in bdrv_co_do_write_zeroes */
+ flags |= BDRV_REQ_MAY_UNMAP;
I'm not sure about this one. I think it is reasonable to expect that
after an explicit write of a buffer filled with zeros the block is
allocated.
In a simple qcow2-on-file case, we basically have three options for
handling all-zero writes:
- Allocate the cluster on a qcow2 and file level and write literal zeros
to it. No metadata updates involved in the next write to the cluster.
- Set the qcow2 zero flag, but leave the allocation in place. The next
write in theory just needs to remove the zero flag (I think in
practice we're doing an unnecessary COW) from the L2 table and that's
it.
- Set the qcow2 zero flag and unmap the cluster on both the qcow2 and
the filesystem layer. The next write causes new allocations in both
layers, which means multiple metadata updates and possibly added
fragmentation. The upside is that we use less disk space if there is
no next write to this cluster.
I think it's pretty clear that the right behaviour depends on your use
case and we can't find an one-size-fits-all solution.