[Libguestfs] Improving supermin appliance startup time (lkvm/qboot)
Richard W.M. Jones
rjones at redhat.com
Tue Mar 15 19:34:32 UTC 2016
On Sun, Mar 13, 2016 at 04:24:53PM +0200, Török Edwin wrote:
> I remembered reading about Intel Clear Containers , and Hyper/Qboot 
> that support startup times measured in the hundreds of miliseconds range.
> On an AMD FX(tm)-8350 'guestfish -a /dev/null run' takes ~4s when run the 2nd/3rd time:
> real 0m4.152s
> user 0m2.120s
> sys 0m0.564s
> Are there any plans on applying these improvements to the supermin appliance?
I was looking at Clear Containers last week.
> I did some quick tests on what would be possible, see below:
> If I try to use virtio 9P instead it becomes a little faster (instead of root fs on block device):
> real 0m3.299s
> user 0m2.084s
> sys 0m1.828s
> To do this we need to load just the virtio and 9p modules:
> let kmods = [
> Modify init.c to mount the 9p root: https://gist.github.com/anonymous/cb32986fd9b85c315ae4
> And wrap qemu to create the 9p device (the mount is just there for a quick test, in reality supermin itself should probably leave the unpacker dir around):
> Next step: kvmtool . Avoids some delays introduced by legacy BIOS boot, and device probing, however
> lkvm currently lacks support for additional virtio-serial channels, so - although the supermin appliance would boot - guestfsd wouldn't be able to communicate with the host:
> I tried QBoot , which requires that you first strip down your kernel+initrd to fit into 8128k,
> and uses an unmodified Qemu otherwise.
> But then I haven't been able to get it to actually launch guestfsd, it hangs somewhere earlier:
> Maybe I removed too many modules here?
> Another possibility would be to run guestfsd with rkt+lkvm stage1, or hyper.sh; but I don't see a way
> to forward a unix socket or a virtio-serial channel through them, would guestfsd be able to work over a TCP socket
> or the main virtio console? (its probably preferable to avoid having any networking in the guestfsd appliance)
>  https://lwn.net/Articles/644675/
>  https://hyper.sh/faq.html
>  https://git.kernel.org/cgit/linux/kernel/git/will/kvmtool.git/tree/README
>  https://github.com/bonzini/qboot
This is all very good analysis.
The issues that I had in brief were:
(1) We could run kvmtool, perhaps by adding a new backend, but it
seems a better idea to add the required features to qemu. Anything
based on kvmtool wouldn't support qcow2 and the myriad other features
of qemu (see also:
(2) On the kernel side, the Intel kernel contains lots of little
hacks, and many baremetal CONFIG_* options are disabled. Hacks can be
upstreamed once we massage them into shape. The real issue is keeping
a single baremetal + virt kernel image, since separate kernel images
would never fly with Fedora or RHEL. That means we cannot just
disable slow drivers/subsystems by having an alternate kernel with
lots of CONFIG_* options disabled, we need to address those problems
(3) DAX / NVDIMM / etc - love them. Not supported upstream (either
kernel or qemu) yet.
(4) Passthrough (eg 9p) filesystems. You touched on that above.
Red Hat doesn't much like 9pfs for several reasons, yet we also don't
have a plausible alternative at the moment. This is mainly a problem
for running fast Docker containers, rather than libguestfs though.
Richard Jones, Virtualization Group, Red Hat http://people.redhat.com/~rjones
Read my programming and virtualization blog: http://rwmj.wordpress.com
Fedora Windows cross-compiler. Compile Windows programs, test, and
build Windows installers. Over 100 libraries supported.
More information about the Libguestfs