[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [PATCH v3 1/2] lvm: add lvm cache logical volume handling
From: |
Daniel Kiper |
Subject: |
Re: [PATCH v3 1/2] lvm: add lvm cache logical volume handling |
Date: |
Fri, 13 Mar 2020 13:23:42 +0100 |
User-agent: |
NeoMutt/20170113 (1.7.2) |
On Wed, Mar 04, 2020 at 02:44:52PM +0800, Michael Chang wrote:
> The lvm cache logical volume is the logical volume consisting of the original
> and the cache pool logical volume. The original is usually on a larger and
> slower storage device while the cache pool is on a smaller and faster one. The
> performance of the original volume can be improved by storing the frequently
> used data on the cache pool to utilize the greater performance of faster
> device.
>
> The default cache mode "writethrough" ensures that any data written will be
> stored both in the cache and on the origin LV, therefore grub can be straight
> to read the original lv as no data loss is guarenteed.
>
> The second cache mode is "writeback", which delays writing from the cache pool
> back to the origin LV to have increased performance. The drawback is potential
> data loss if losing the associated cache device.
>
> During the boot time grub reads the LVM offline i.e. LVM volumes are not
> activated and mounted, hence it should be fine to read directly from original
> lv since all cached data should have been flushed back in the process of
> taking
> it offline.
>
> It is also not much helpful to the situation by adding fsync calls to the
> install code. The fsync did not force to write back dirty cache to the
> original
> device and rather it would update associated cache metadata to complete the
> write transaction with the cache device. IOW the writes to cached blocks still
> go only to the cache device.
>
> To write back dirty cache, as lvm cache did not support dirty cache flush per
> block range, there'no way to do it for file. On the other hand the "cleaner"
> policy is implemented and can be used to write back "all" dirty blocks in a
> cache, which effectively drain all dirty cache gradually to attain and last in
> the "clean" state, which can be useful for shrinking or decommissioning a
> cache. The result and effect is not what we are looking for here.
>
> In conclusion, as it seems no way to enforce file writes to the original
> device, grub may suffer from power failure as it cannot assemble the cache
> device and read the dirty data from it. However since the case is only
> applicable to writeback mode which is sensitive to data lost in nature, I'd
> still like to propose my (relatively simple) patch and treat reading dirty
> cache as improvement.
>
> Signed-off-by: Michael Chang <address@hidden>
> ---
>
> Changes since v3:
> * Add some issue discussed in review process to commit message
> * Add note on LVM cache booting to grub manual
> * Some coding style change
> * Fix some problem in error handling
>
> Changes since v2:
> * Move struct cache_lv definition to the beginning of file
> * Add handling when grub_(zalloc|malloc|strdup) etc failed
>
> grub-core/disk/lvm.c | 185
> +++++++++++++++++++++++++++++++++++++++++++++++++++
> 1 file changed, 185 insertions(+)
>
> diff --git a/grub-core/disk/lvm.c b/grub-core/disk/lvm.c
> index 0cbd0dd16..e97f0a1f0 100644
> --- a/grub-core/disk/lvm.c
> +++ b/grub-core/disk/lvm.c
> @@ -33,6 +33,14 @@
>
> GRUB_MOD_LICENSE ("GPLv3+");
>
> +struct cache_lv
> +{
> + struct grub_diskfilter_lv *lv;
> + char *cache_pool;
> + char *origin;
> + struct cache_lv *next;
> +};
> +
>
> /* Go the string STR and return the number after STR. *P will point
> at the number. In case STR is not found, *P will be NULL and the
> @@ -95,6 +103,34 @@ grub_lvm_check_flag (const char *p, const char *str,
> const char *flag)
> }
> }
>
> +static void
> +grub_lvm_free_cache_lvs (struct cache_lv *cache_lvs)
> +{
> + struct cache_lv *cache;
> +
> + while ((cache = cache_lvs))
> + {
> + cache_lvs = cache_lvs->next;
> +
> + if (cache->lv)
> + {
> + unsigned int i;
> +
> + for (i = 0; i < cache->lv->segment_count; ++i)
> + if (cache->lv->segments)
> + grub_free (cache->lv->segments[i].nodes);
> + grub_free (cache->lv->segments);
> + grub_free (cache->lv->fullname);
> + grub_free (cache->lv->idname);
> + grub_free (cache->lv->name);
> + }
> + grub_free (cache->lv);
> + grub_free (cache->origin);
> + grub_free (cache->cache_pool);
> + grub_free (cache);
> + }
> +}
> +
> static struct grub_diskfilter_vg *
> grub_lvm_detect (grub_disk_t disk,
> struct grub_diskfilter_pv_id *id,
> @@ -243,6 +279,8 @@ grub_lvm_detect (grub_disk_t disk,
>
> if (! vg)
> {
> + struct cache_lv *cache_lvs = NULL;
> +
> /* First time we see this volume group. We've to create the
> whole volume group structure. */
> vg = grub_malloc (sizeof (*vg));
> @@ -672,6 +710,101 @@ grub_lvm_detect (grub_disk_t disk,
> seg->nodes[seg->node_count - 1].name = tmp;
> }
> }
> + else if (grub_memcmp (p, "cache\"",
> + sizeof ("cache\"") - 1) == 0)
> + {
> + struct cache_lv *cache = NULL;
> +
> + char *p2, *p3;
> + grub_ssize_t sz;
Why not grub_size_t?
> +
> + cache = grub_zalloc (sizeof (*cache));
> + if (!cache)
> + goto cache_lv_fail;
> + cache->lv = grub_zalloc (sizeof (*cache->lv));
> + if (!cache->lv)
> + goto cache_lv_fail;
> + grub_memcpy (cache->lv, lv, sizeof (*cache->lv));
> +
> + if (lv->fullname)
> + {
> + cache->lv->fullname = grub_strdup (lv->fullname);
> + if (!cache->lv->fullname)
> + goto cache_lv_fail;
> + }
> + if (lv->idname)
> + {
> + cache->lv->idname = grub_strdup (lv->idname);
> + if (!cache->lv->idname)
> + goto cache_lv_fail;
> + }
> + if (lv->name)
> + {
> + cache->lv->name = grub_strdup (lv->name);
> + if (!cache->lv->name)
> + goto cache_lv_fail;
> + }
> +
> + skip_lv = 1;
> +
> + p2 = grub_strstr (p, "cache_pool = \"");
> + if (!p2)
> + goto cache_lv_fail;
> +
> + p2 = grub_strchr (p2, '"');
> + if (!p2)
> + goto cache_lv_fail;
> +
> + p3 = ++p2;
> + p3 = grub_strchr (p3, '"');
> + if (!p3)
> + goto cache_lv_fail;
> +
> + sz = p3 - p2;
> +
> + cache->cache_pool = grub_malloc (sz + 1);
> + if (!cache->cache_pool)
> + goto cache_lv_fail;
> + grub_memcpy (cache->cache_pool, p2, sz);
> + cache->cache_pool[sz] = '\0';
> +
> + p2 = grub_strstr (p, "origin = \"");
> + if (!p2)
> + goto cache_lv_fail;
> +
> + p2 = grub_strchr (p2, '"');
> + if (!p2)
> + goto cache_lv_fail;
> +
> + p3 = ++p2;
> + p3 = grub_strchr (p3, '"');
> + if (!p3)
> + goto cache_lv_fail;
> +
> + sz = p3 - p2;
> +
> + cache->origin = grub_malloc (sz + 1);
> + if (!cache->origin)
> + goto cache_lv_fail;
> + grub_memcpy (cache->origin, p2, sz);
> + cache->origin[sz] = '\0';
> +
> + cache->next = cache_lvs;
> + cache_lvs = cache;
> + break;
> +
> + cache_lv_fail:
> + if (cache)
> + {
> + grub_free (cache->origin);
> + grub_free (cache->cache_pool);
> + grub_free (cache->lv->fullname);
If "cache = grub_zalloc (sizeof (*cache));" fails above then
here GRUB crashes due to NULL pointer dereference...
> + grub_free (cache->lv->idname);
...this has to be fixed too...
> + grub_free (cache->lv->name);
Ditto...
> + grub_free (cache);
> + }
> + goto fail4;
> + }
> else
> {
> #ifdef GRUB_UTIL
Daniel