[vfio-users] vfio fails Guest FreeBSD9.3 host Fedora 23
chintu hetam
rometoroam at gmail.com
Tue Jun 7 14:15:11 UTC 2016
1G hugepages are allocated with grub at boot time.
[root at localhost vcr]# cat /etc/default/grub
GRUB_TIMEOUT=5
GRUB_DISTRIBUTOR="$(sed 's, release .*$,,g' /etc/system-release)"
GRUB_DEFAULT=saved
GRUB_DISABLE_SUBMENU=true
GRUB_TERMINAL_OUTPUT="console"
GRUB_CMDLINE_LINUX="rd.lvm.lv=fedora/root rd.lvm.lv=fedora/swap
hugepagesz=1G hugepages=32 iommu=pt intel_iommu=on rhgb quiet"
GRUB_DISABLE_RECOVERY="true"
and after doing grub2-mkconfig boot file looks like
if [ x$feature_platform_search_hint = xy ]; then
search --no-floppy --fs-uuid --set=root --hint-bios=hd0,msdos1
--hint-efi=hd0,msdos1 --hint-baremetal=ahci0,msdos1 --hint='hd0,msdos1'
92f30a97-3f52-4352-a28d-ce0ba665377d
else
search --no-floppy --fs-uuid --set=root
92f30a97-3f52-4352-a28d-ce0ba665377d
fi
linux16 /vmlinuz-4.4.6-300.fc23.x86_64 root=/dev/mapper/fedora-root
ro rd.lvm.lv=fedora/root rd.lvm.lv=fedora/swap hugepagesz=1G hugepages=32
iommu=pt intel_iommu=on rhgb quiet
As there's isn't any 1G default huge page size in grub command line, it
defaults to 2M huge pages
- cat /proc/meminfo
VmallocUsed: 0 kB
VmallocChunk: 0 kB
HardwareCorrupted: 0 kB
AnonHugePages: 131072 kB
CmaTotal: 0 kB
CmaFree: 0 kB
HugePages_Total: 65536
HugePages_Free: 32768
HugePages_Rsvd: 32768
HugePages_Surp: 0
Hugepagesize: 2048 kB
DirectMap4k: 1601216 kB
DirectMap2M: 72767488 kB
DirectMap1G: 330301440 kB
[root at localhost vcr]# cat
/sys/devices/system/node/node0/hugepages/hugepages-2048kB/nr_hugepages
32768
[root at localhost vcr]# cat
/sys/devices/system/node/node0/hugepages/hugepages-1048576kB/nr_hugepages
16
[root at localhost vcr]# cat
/sys/devices/system/node/node1/hugepages/hugepages-1048576kB/nr_hugepages
16
[root at localhost vcr]# cat
/sys/devices/system/node/node1/hugepages/hugepages-2048kB/nr_hugepages
32768
i set 2M hugepages with following command respectively
echo 32768 >
/sys/devices/system/node/node0/hugepages/hugepages-2048kB/nr_hugepages
echo 32768 >
/sys/devices/system/node/node1/hugepages/hugepages-2048kB/nr_hugepages
will update here after trying your steps..
-chintu-
On Wed, Jun 1, 2016 at 6:56 PM, Alex Williamson <alex.williamson at redhat.com>
wrote:
> On Wed, 1 Jun 2016 15:46:10 -0400
> chintu hetam <rometoroam at gmail.com> wrote:
>
> > hugepages doesn't work with VFIO it works with vhost.
>
> TBH I don't have a lot of faith in your setup since you haven't really
> shown how you're allocating 1G hugepages, or even 2M, which many people
> here are using successfully. Here's a simple test:
>
> # cd /sys/kernel/mm/hugepages/hugepages-1048576kB/
> # cat nr_hugepages
>
> How many do you have available? If less than 32, then:
>
> # echo 32 > nr_hugepages
>
> Recheck how many are available, you may not be able to get 32 except
> via boot options. If less, adjust the VM memory size accordingly in
> the command below.
>
> Mount your 1G hugepages:
>
> # mkdir /hugepages1G
> # mount -t hugetlbfs -o pagesize=1G hugetlbfs /hugepages1G
>
> Bind your device to vfio-pci:
>
> # virsh nodedev-detach pci_0000_aa_06_0
>
> Start a simple VM:
>
> # /usr/bin/qemu-kvm -m 32G -mem-path /hugepages1G -nodefaults -monitor
> stdio -nographic -device vfio-pci,host=aa:06.0
>
> This will give you a (qemu) prompt where you can type 'quit' or ^C to
> kill it. If this works, then your problem is with configuring your
> system to use 1G hugepages or with libvirt. Please test. Thanks,
>
> Alex
>
--
-chintu-
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/vfio-users/attachments/20160607/53966de4/attachment.htm>
More information about the vfio-users
mailing list