qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-devel] [RFC PATCH] vfio/pci: Add ioeventfd support


From: Peter Xu
Subject: Re: [Qemu-devel] [RFC PATCH] vfio/pci: Add ioeventfd support
Date: Sun, 11 Feb 2018 11:09:36 +0800
User-agent: Mutt/1.9.1 (2017-09-22)

On Fri, Feb 09, 2018 at 02:45:41PM -0700, Alex Williamson wrote:
> On Fri, 9 Feb 2018 15:05:11 +0800
> Peter Xu <address@hidden> wrote:
> 
> > On Tue, Feb 06, 2018 at 05:08:14PM -0700, Alex Williamson wrote:
> > 
> > [...]
> > 
> > > +long vfio_pci_ioeventfd(struct vfio_pci_device *vdev, loff_t offset,
> > > +                 uint64_t data, int count, int fd)
> > > +{
> > > + struct pci_dev *pdev = vdev->pdev;
> > > + loff_t pos = offset & VFIO_PCI_OFFSET_MASK;
> > > + int ret, bar = VFIO_PCI_OFFSET_TO_INDEX(offset);
> > > + struct vfio_pci_ioeventfd *ioeventfd;
> > > + int (*handler)(void *, void *);
> > > + unsigned long val;
> > > +
> > > + /* Only support ioeventfds into BARs */
> > > + if (bar > VFIO_PCI_BAR5_REGION_INDEX)
> > > +         return -EINVAL;
> > > +
> > > + if (pos + count > pci_resource_len(pdev, bar))
> > > +         return -EINVAL;
> > > +
> > > + /* Disallow ioeventfds working around MSI-X table writes */
> > > + if (bar == vdev->msix_bar &&
> > > +     !(pos + count <= vdev->msix_offset ||
> > > +       pos >= vdev->msix_offset + vdev->msix_size))
> > > +         return -EINVAL;
> > > +
> > > + switch (count) {
> > > + case 1:
> > > +         handler = &vfio_pci_ioeventfd_handler8;
> > > +         val = data;
> > > +         break;
> > > + case 2:
> > > +         handler = &vfio_pci_ioeventfd_handler16;
> > > +         val = le16_to_cpu(data);
> > > +         break;
> > > + case 4:
> > > +         handler = &vfio_pci_ioeventfd_handler32;
> > > +         val = le32_to_cpu(data);
> > > +         break;
> > > +#ifdef iowrite64
> > > + case 8:
> > > +         handler = &vfio_pci_ioeventfd_handler64;
> > > +         val = le64_to_cpu(data);
> > > +         break;
> > > +#endif
> > > + default:
> > > +         return -EINVAL;
> > > + }
> > > +
> > > + ret = vfio_pci_setup_barmap(vdev, bar);
> > > + if (ret)
> > > +         return ret;
> > > +
> > > + mutex_lock(&vdev->ioeventfds_lock);
> > > +
> > > + list_for_each_entry(ioeventfd, &vdev->ioeventfds_list, next) {
> > > +         if (ioeventfd->pos == pos && ioeventfd->bar == bar &&
> > > +             ioeventfd->data == data && ioeventfd->count == count) {
> > > +                 if (fd == -1) {
> > > +                         vfio_virqfd_disable(&ioeventfd->virqfd);
> > > +                         list_del(&ioeventfd->next);
> > > +                         kfree(ioeventfd);
> > > +                         ret = 0;
> > > +                 } else
> > > +                         ret = -EEXIST;
> > > +
> > > +                 goto out_unlock;
> > > +         }
> > > + }
> > > +
> > > + if (fd < 0) {
> > > +         ret = -ENODEV;
> > > +         goto out_unlock;
> > > + }
> > > +
> > > + ioeventfd = kzalloc(sizeof(*ioeventfd), GFP_KERNEL);
> > > + if (!ioeventfd) {
> > > +         ret = -ENOMEM;
> > > +         goto out_unlock;
> > > + }
> > > +
> > > + ioeventfd->pos = pos;
> > > + ioeventfd->bar = bar;
> > > + ioeventfd->data = data;
> > > + ioeventfd->count = count;
> > > +
> > > + ret = vfio_virqfd_enable(vdev->barmap[ioeventfd->bar] + ioeventfd->pos,
> > > +                          handler, NULL, (void *)val,
> > > +                          &ioeventfd->virqfd, fd);
> > > + if (ret) {
> > > +         kfree(ioeventfd);
> > > +         goto out_unlock;
> > > + }
> > > +
> > > + list_add(&ioeventfd->next, &vdev->ioeventfds_list);  
> > 
> > Is there a limit on how many ioeventfds that can be created?
> > 
> > IIUC we'll create this eventfd "automatically" if a MMIO addr/data
> > triggered continuously for N=10 times, then would it be safer we have
> > a limitation on maximum eventfds?  Or not sure whether a malicious
> > guest can consume the host memory by sending:
> > 
> > - addr1/data1, 10 times
> > - addr2/data2, 10 times
> > - ...
> > 
> > To create unlimited ioeventfds?  Thanks,
> 
> Good question, it is somewhat exploitable in the guest the way it's
> written, however a user process does have an open file limit and each
> eventfd consumes a file handle, so unless someone is running QEMU with
> unlimited file handles, there is a built-in limit.  Two problems remain
> though:
> 
> First, is it still a bad idea that a user within a guest can target
> this device page to consume all of the QEMU process' open file handles,
> even if ultimately they're only harming themselves?  What would a
> reasonable cap of file descriptors for this purpose be?  How would we
> know which are actively used and which could be expired?  Currently
> only 2 are registered, the MSI-ACK address and some unknown secondary
> one that's low frequency, but enough to trigger the algorithm here (and
> doesn't seem harmful to let it get enabled).  We could therefore
> arbitrarily pick 5 as an upper limit here, maybe with a warning if the
> code hits that limit.
> 
> Second, is there still an exploit in the proposed vfio interface that a
> user could re-use a single file descriptor for multiple vfio
> ioeventfds.  I don't know.  I thought about looking to see whether a
> file descriptor is re-used, but then I wondered if that might actually
> be kind of a neat and potentially useful interface that a single
> eventfd could trigger multiple device writes.

Do you know any of possible scenario for this?  Since that sounds like
an interesting idea.

> It looks like KVM
> arbitrarily sets a limit of 1000 iobus devices, where each ioeventfd
> would be such a device.  Perhaps we take the same approach and pick an
> arbitrary high upper limit per vfio device.  What's your opinion?

I did a "ulimit -n" and I got 1024 as default.  So even if someone
granted the QEMU process with more/unlimited file handles, we'll still
have a hard limit of ~1000 vfio-ioeventfds, which sounds safe.

But I agree that we can still have a even smaller limit for vfio,
especially we know explicitly on how many we may need in most cases.
I'll follow your choice of number because I totally have no good idea
on that.  While if we are going to have a limitation, I'm thinking
whether it'll be good we have a WARN_ONCE when that limit is reached.

Thanks,

-- 
Peter Xu



reply via email to

[Prev in Thread] Current Thread [Next in Thread]