qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH V4 02/19] physmem: fd-based shared memory


From: Peter Xu
Subject: Re: [PATCH V4 02/19] physmem: fd-based shared memory
Date: Mon, 16 Dec 2024 13:19:34 -0500

On Fri, Dec 13, 2024 at 11:41:45AM -0500, Steven Sistare wrote:
> On 12/12/2024 4:22 PM, Peter Xu wrote:
> > On Thu, Dec 12, 2024 at 03:38:00PM -0500, Steven Sistare wrote:
> > > On 12/9/2024 2:42 PM, Peter Xu wrote:
> > > > On Mon, Dec 02, 2024 at 05:19:54AM -0800, Steve Sistare wrote:
> > > > > @@ -2089,13 +2154,23 @@ RAMBlock *qemu_ram_alloc_internal(ram_addr_t 
> > > > > size, ram_addr_t max_size,
> > > > >        new_block->page_size = qemu_real_host_page_size();
> > > > >        new_block->host = host;
> > > > >        new_block->flags = ram_flags;
> > > > > +
> > > > > +    if (!host && !xen_enabled()) {
> > > > 
> > > > Adding one more xen check is unnecessary.  This patch needed it could 
> > > > mean
> > > > that the patch can be refactored.. because we have xen checks in both
> > > > ram_block_add() and also in the fd allocation path.
> > > > 
> > > > At the meantime, see:
> > > > 
> > > > qemu_ram_alloc_from_fd():
> > > >       if (kvm_enabled() && !kvm_has_sync_mmu()) {
> > > >           error_setg(errp,
> > > >                      "host lacks kvm mmu notifiers, -mem-path 
> > > > unsupported");
> > > >           return NULL;
> > > >       }
> > > > 
> > > > I don't think any decent kernel could hit this, but that could be 
> > > > another
> > > > sign that this patch duplicated some file allocations.
> > > > 
> > > > > +        if ((new_block->flags & RAM_SHARED) &&
> > > > > +            !qemu_ram_alloc_shared(new_block, &local_err)) {
> > > > > +            goto err;
> > > > > +        }
> > > > > +    }
> > > > > +
> > > > >        ram_block_add(new_block, &local_err);
> > > > > -    if (local_err) {
> > > > > -        g_free(new_block);
> > > > > -        error_propagate(errp, local_err);
> > > > > -        return NULL;
> > > > > +    if (!local_err) {
> > > > > +        return new_block;
> > > > >        }
> > > > > -    return new_block;
> > > > > +
> > > > > +err:
> > > > > +    g_free(new_block);
> > > > > +    error_propagate(errp, local_err);
> > > > > +    return NULL;
> > > > >    }
> > > > 
> > > > IIUC we only need to conditionally convert an anon-allocation into an
> > > > fd-allocation, and then we don't need to mostly duplicate
> > > > qemu_ram_alloc_from_fd(), instead we reuse it.
> > > > 
> > > > I do have a few other comments elsewhere, but when I was trying to 
> > > > comment.
> > > > E.g., we either shouldn't need to bother caching qemu_memfd_check()
> > > > results, or do it in qemu_memfd_check() directly.. and some more.
> > > 
> > > Someone thought it a good idea to cache the result of 
> > > qemu_memfd_alloc_check,
> > > and qemu_memfd_check will be called more often.  I'll cache the result 
> > > inside
> > > qemu_memfd_check for the special case of flags=0.
> > 
> > OK.
> > 
> > > 
> > > > Then I think it's easier I provide a patch, and also show that it can be
> > > > also smaller changes to do the same thing, with everything fixed up
> > > > (e.g. addressing above mmu notifier missing issue).  What do you think 
> > > > as
> > > > below?
> > > 
> > > The key change you make is calling qemu_ram_alloc_from_fd instead of 
> > > file_ram_alloc,
> > > which buys the xen and kvm checks for free.  Sounds good, I will do that 
> > > in the
> > > context of my patch.
> > > 
> > > Here are some other changes in your patch, and my responses:
> > > 
> > > I will drop the "Retrying using MAP_ANON|MAP_SHARED" message, as you did.
> > > 
> > > However, I am keeping QEMU_VMALLOC_ALIGN, qemu_set_cloexec, and 
> > > trace_qemu_ram_alloc_shared.
> > 
> > I guess no huge deal on these, however since we're talking..  Is that
> > QEMU_VMALLOC_ALIGN from qemu_anon_ram_alloc()?
> > 
> > A quick dig tells me that it was used to be for anon THPs..
> > 
> >      commit 36b586284e678da28df3af9fd0907d2b16f9311c
> >      Author: Avi Kivity <avi@redhat.com>
> >      Date:   Mon Sep 5 11:07:05 2011 +0300
> > 
> >      qemu_vmalloc: align properly for transparent hugepages and KVM
> > 
> > And I'm guessing if at that time was also majorly for guest ram.
> > 
> > Considering that this path won't make an effect until the new aux mem
> > option is on, I'd think it better to stick without anything special like
> > QEMU_VMALLOC_ALIGN, until it's justified to be worthwhile.  E.g., Avi used
> > to explicitly mention this in that commit message:
> > 
> >      Adjust qemu_vmalloc() to honor that requirement.  Ignore it for small 
> > regions
> >      to avoid fragmentation.
> > 
> > And this is exactly mostly small regions when it's AUX.. probably except
> > VGA, but it'll be SHARED on top of shmem not PRIVATE on anon anyway... so
> > it'll be totally different things.
> > 
> > So I won't worry on that 2M alignment, and I will try to not carry over
> > that, because then trying to remove it will be harder.. even when we want.
> 
> Yes, currently the aux allocations get QEMU_VMALLOC_ALIGN alignment in
> qemu_anon_ram_alloc.  I do the same for the shared fd mappings to guarantee
> no performance regression,

I don't know how we could guarantee that at all - anon and shmem uses
different knobs to enable/disable THPs after all.. For example:

  $ ls /sys/kernel/mm/transparent_hugepage/*enabled
  /sys/kernel/mm/transparent_hugepage/enabled
  /sys/kernel/mm/transparent_hugepage/shmem_enabled

And their default values normally differ too... it means after switching to
fd based we do face the possibility that thp can be gone at least on the
1st 2mb.

When I was suggesting it, I was hoping thp doesn't really matter that lot
on aux mem, even for VGA.

Btw, I don't even think the alignment will affect THP allocations for the
whole vma, anyway?  I mean, it's only about the initial 2MB portion.. IOW,
when not aligned, I think the worst case is we have <2MB at start address
that is not using THP, but later on when it starts to align with 2MB, THPs
will be allocated again.

The challenge is more on the "fd-based" side, where shmem on most distros
will disable THP completely.

> as some of them are larger than 2M and would
> benefit from using huge pages.  The VA fragmentation is trivial for this small
> number of aux blocks in a 64-bit address space, and is no different than it 
> was
> for qemu_anon_ram_alloc.
> 
> > For the 2nd.. Any quick answer on why explicit qemu_set_cloexec() needed?
> 
> qemu sets cloexec for all descriptors it opens to prevent them from 
> accidentally
> being leaked to another process via fork+exec.

But my question is why this is special?  For example, we don't do that for
"-object memory-backend-memfd", am I right?

> 
> > For 3rd, tracepoint would definitely be fine whenever you feel necessary.
> > 
> > > Also, when qemu_memfd_create + qemu_ram_alloc_from_fd fails, qemu should 
> > > fail and exit,
> > > and not fall back, because something unexpected went wrong.  David said 
> > > the same.
> > 
> > Why?  I was trying to rely on such fallback to make it work on e.g. Xen.
> > In that case, Xen fails there and fallback to xen_ram_alloc() inside the
> > later call to ram_block_add(), no?
> 
> Why -- because something went wrong that should have worked, and we should 
> report the
> first fault so its cause can be fixed and cpr can be used.

Ahh so it's only about the corner cases where CPR could raise an error?
Can we rely on the failure later on "migrate" command to tell which
ramblock doesn't support it, so the user could be aware as well?

> 
> However, to do the above, but still quietly fallback if qemu_ram_alloc_from_fd
> fails because of xen or kvm, I would need to return different error codes from
> qemu_ram_alloc_from_fd.  Doable, but requires tweaks to all occurrences of
> qemu_ram_alloc_from_fd.
> 
> And BTW, qemu_ram_alloc_from_fd is defined for CONFIG_POSIX only.  I need
> to modify the call site in the patch accordingly.

Yep, I was thinking maybe qemu_ram_alloc_from_fd() had a stub function,
indeed looks not..  "allocating the fd" part definitely has, which I
remember I checked..

> 
> Overall, I am not convinced that using qemu_ram_alloc_from_fd in this patch
> is better/simpler than my V4 patch using file_ram_alloc, plus adding xen and
> kvm_has_sync_mmu checks in qemu_ram_alloc_internal.

As long as you don't need to duplicate these two checks (or duplicate any
such check..) I'm ok.

Reusing qemu_ram_alloc_from_fd() still sounds like the easiest to go.  Yes
we'll need to teach it about resize(), used_length etc. to it, but they all
look sane to me.  We didn't have those simply because we don't have use of
them, now we want to have resizable fd-based mem, that's the right thing to
do to support that on fd allocations.

OTOH, duplicating xen/mmu checks isn't sane to me.. :( It will make the
code harder to maintain because the 3rd qemu_ram_alloc_from_fd() in the
future will need to duplicate it once more (or worse, forget it again until
xen / old kernels reports a failure)..

-- 
Peter Xu




reply via email to

[Prev in Thread] Current Thread [Next in Thread]