qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Qemu-devel] Re: [PATCHv2 09/12] vhost: vhost net support


From: Anthony Liguori
Subject: [Qemu-devel] Re: [PATCHv2 09/12] vhost: vhost net support
Date: Sun, 28 Feb 2010 10:02:29 -0600
User-agent: Mozilla/5.0 (X11; U; Linux x86_64; en-US; rv:1.9.1.5) Gecko/20091209 Fedora/3.0-4.fc12 Lightning/1.0pre Thunderbird/3.0

On 02/27/2010 01:38 PM, Michael S. Tsirkin wrote:
On Fri, Feb 26, 2010 at 09:18:03AM -0600, Anthony Liguori wrote:
On 02/26/2010 08:49 AM, Michael S. Tsirkin wrote:
KVM code needs all kind of work-arounds for KVM specific issues.
It also assumes that KVM is registered at startup, so it
does not try to optimize finding slots.

No, the slot mapping changes dynamically so KVM certainly needs to
optimize this.
Maybe, but it does not, KVM algorithms are n^2 or worse.

But n is small and the mappings don't change frequently.

More importantly, they change at the exact same times for vhost as they do for kvm. So even if vhost has an O(n) algorithm, the KVM code gets executed either immediately before or immediately after the vhost code so your optimizations are lost in KVM's O(n^2) algorithm.

Mine has no bugs, let's switch to it!

Seriously, need to tread very carefully here.
This is why I say: merge it, then look at how to reuse code.

Once it's merged, there's no incentive to look at reusing code.
Again, I don't think this is a huge burden to vhost.  The two bits of code
literally do exactly the same thing.  They just use different data
structures that ultimately contain the same values.
Not exactly. For example kvm track ROM and video ram addresses.

KVM treats ROM and RAM the same (it even maps ROM as RAM). There is no special handling for video ram addresses.

There is some magic in the VGA code to switch the VGA LFB from mmio to ram when possible but that happens at a higher layer.

'++i' is an odd thing to do in C in a for() loop.  We're not explicit
about it in Coding Style but the vast majority of code just does
'i++'.
Ugh. Do we really need to specify every little thing?

I don't care that much about coding style. I don't care if there are curly brackets on single line ifs.

However, it's been made very clear to me that most other people do and that it's something that's important to enforce.

Hmm. I'll look into it.
I actually think that for functions that just do a list of things
unconditionally, without branches or loops, or with just error handling
as here, it is perfectly fine for them to be of any length.

Like I said, just a suggestion.

You never unmap() the mapped memory and you're cheating by assuming that
the virtio rings have a constant mapping for the life time of a guest.
That's not technically true.  My concern is that since a guest can
trigger remappings (by adjusting PCI mappings) badness can ensue.

I do not know how this can happen. What do PCI mappings have to do with this?
Please explain. If it can, vhost will need notification to update.

If a guest modifies the bar for an MMIO region such that it happens to
exist in RAM, while this is a bad thing for the guest to do, I don't
think we do anything to stop it.  When the region gets remapped, the
result will be that the mapping will change.
So IMO this is the bug. If there's a BAR that matches RAM
physical address, it should never get mapped. Any idea how
to check this?

We could check it when the BAR is mapped in the PCI layer. I'm suspicious there are other ways a guest can enforce/determine mappings though.

Generally speaking, I think it's necessary to assume that a guest can manipulate memory mappings. If we can prove that a guest cannot, it would definitely simplify the code a lot. I'd love to make the same assumptions in virtio userspace before it's actually a big source of overhead.

I'm pretty sure though that we have to let a guest control mappings though.

Within qemu, because we carry the qemu_mutex, we know that the mappings
are fixed as long as we're in qemu.  We're very careful to assume that
we don't rely on a mapping past when we drop the qemu_mutex.

With vhost, you register a slot table and update it whenever mappings
change.  I think that's good enough for dealing with ram addresses.  But
you pass the virtual address for the rings and assume those mappings
never change.
So, the issue IMO is that an MMIO address gets passed instead of RAM.
There's no reason to put virtio rings not in RAM, we just need to
verify this.

Yes, but we don't always map PCI IO regions as MMIO or PIO. In particular, for VGA devices (particularly VMware VGA), we map certain IO regions as RAM because that's how the device is designed. Likewise, if we do shared memory PCI devices using IO regions as the ram contents, we would be mapping those as ram too.

So just checking to see if the virtio ring area is RAM or not is not enough. A guest may do something that causes a virtio ring to still be ram, but a different ram address. Now the vhost code is writing to RAM that it thinks is physical address X but is really guest physical address Y.

This is not something that a guest can use to break into qemu, but it is an emulation bug and depending on the guest OS, it may be possible to use it to do a privilege escalation within the guest.

I think the only way to handle this is to explicitly check for changes in the physical addresses the rings are mapped at and do the appropriate ioctls to vhost to let it know if the ring's address has changed.

I'm pretty sure a guest can cause those to change and I'm not 100% sure,
but I think it's a potential source of exploits if you assume a mapping.
In the very least, a guest can trick vhost into writing to ram that it
wouldn't normally write to.
This seems harmless. guest can write anywhere in ram, anyway.

Not all guest code is created equal and if we're writing to the wrong guest ram location, it can potentially circumvent the guest's security architecture.

Regards,

Anthony Liguori





reply via email to

[Prev in Thread] Current Thread [Next in Thread]