[vfio-users] Q on vfio-pci driver usage on Host

Alex Williamson alex.williamson at redhat.com
Mon Apr 13 18:39:40 UTC 2020


On Mon, 13 Apr 2020 10:33:21 -0700
Ravi Kerur <rkerur at gmail.com> wrote:

> On Mon, Apr 13, 2020 at 8:36 AM Alex Williamson <alex.williamson at redhat.com>
> wrote:
> 
> > On Sun, 12 Apr 2020 09:10:49 -0700
> > Ravi Kerur <rkerur at gmail.com> wrote:
> >  
> > > Hi,
> > >
> > > I use Intel NICs for PF and VF devices. VFs are assigned to virtual
> > > machines and PF is used on the Host. I have intel-iommu=on on GRUB which
> > > enables DMAR and IOMMU capabilities (checked via 'dmesg | grep -e IOMMU  
> > -e  
> > > DMAR) and use DPDK for datapath acceleration.
> > >
> > > Couple of clarifications I need in terms of vfio-pci driver usage
> > >
> > > (1) with intel-iommu=pt (Passthrough mode), PF device on host can bind to
> > > either igb_uio or vfio-pci driver and similarly VF devices assigned to  
> > each  
> > > VM can bind to either igb_uio or vfio-pci driver via Qemu  
> >
> > Note that the actual option is 'intel_iommu=on iommu=pt'.
> >  
> 
> My mistake,
> 
> >  
> > > (2) with intel-iommu=on (IOMMU enabled), PF device on host must bind to
> > > vfio-pci driver and similarly VF devices assigned to each VM much bind to
> > > vfio-pci driver. When IOMMU is enabled, only vfio-pci should be used?  
> >
> > When an IOMMU is present, we refer to the address space through which a
> > device performs DMA as the I/O Virtual Address space, or IOVA.  When
> > the IOMMU is in passthrough mode, we effectively create an identity
> > mapping of physical addresses through the IOVA space.  Therefore to
> > program a device to perform a DMA to user memory, the user only needs
> > to perform a virtual to physical translation on the address and the
> > device can DMA directly with that physical address thanks to the
> > identity map.  When we're not in passthrough mode, we need to actually
> > create a mapping through the IOMMU to allow the device to access that
> > physical memory.  VFIO is the only userspace driver interface that I'm
> > aware of that provides this latter functionality.  Therefore, yes, if
> > you have the IOMMU enabled and not in passthrough mode, your userspace
> > driver needs support for programming the IOMMU, which vfio-pci provides.
> >
> > Also, having both the PF and VFs owned by userspace drivers presents
> > some additional risks, for example the PF may have access to the data
> > accessed by the VF, or at least be able to deny service to the VF.
> > There have been various hacks around this presented by the DPDK
> > community, essentially enabling SR-IOV underneath vfio-pci, without the
> > driver's knowledge.  These are very much dis-recommended, IMO.
> > However, we have added SR-IOV support to the vfio-pci driver in kernel
> > v5.7 and DPDK support is under development, which represents this trust
> > and collaboration between PF and VF drivers using a new VF token
> > concept.  I'd encourage you to look for this if your configuration does
> > require both PF and VF drivers in userspace.  A much more normal
> > configuration to this point has been that the PF makes use of a host
> > driver (ex. igb, ixgbe, i40e, etc.) while the VF is bound to vfio-pci
> > for userspace drivers.  In this configuration the host driver is
> > considered trusted and we don't need to invent new mechanisms to
> > indicate collaboration between userspace drivers.  Thanks,
> >  
> 
> Thanks for the information. Clearly understand what I need to do. Where can
> I find information on vfio-pci sr-iov support (writeup/design)?

https://lore.kernel.org/kvm/158396044753.5601.14804870681174789709.stgit@gimli.home/

http://mails.dpdk.org/archives/dev/2020-April/163095.html

Thanks,
Alex




More information about the vfio-users mailing list