[libvirt] Live migration sanity checks

Andrew Cathrow acathrow at redhat.com
Tue Oct 28 14:21:11 UTC 2008


On Wed, 2008-09-03 at 00:13 +0100, Daniel P. Berrange wrote:

> On Tue, Sep 02, 2008 at 07:38:59AM +0200, Chris Lalancette wrote:
> > Richard W.M. Jones wrote:
> > > On Fri, Aug 29, 2008 at 10:44:50AM +0200, Chris Lalancette wrote:
> > >> Things that I've missed?
> > > 
> > > Maybe a good place for this list is on the wiki?  On the actual
> > > feature/todo page.
> > > 
> > > I'd like to see where KVM is going to go with this, since it seems
> > > they are going to implement migration checking.
> > 
> > Yes, OK, I've put that there now.  I also want to see what KVM does here;
> > however, I don't think that prevents us from implementing our own, since we
> > would still need similar things for other hypervisors (Xen, etc.).
> 
> Following on from this, there are some things it is simply not practical
> for the underlying hypervisor to check for itself - specifically things 
> that require knowledge outside the scope of the HV. For example, how
> would the hypervisor ever know whether /dev/sda1 on the source box was
> the same as /dev/sda1 on the destination box. This information is only
> available at a higher level. Indeed for some of this, even libvirt can
> not answer the question, and oVirt would have to make decisions directly.

Does mandating the use of lablels or UUID solve the disk naming problem?
Or does the storage pool model bypass this altogether?

> 
> Looking at Chris' list of things to check for I think one thing is very
> clear - a simple boolean test is not a useful API model at the  libvirt,
> let alone the hypervisor level. There are a series of items that need to
> be checked. Some may appear have a straight yes/no answer, but in fact 
> the eventual decision is a matter of application policy.
> 
> For example it may seem 'obvious' that you cannot migrate a i386 guest
> from an x86_64 host onto a PPC64 host. This would be a bad assumption 
> though, because you may be quite happy to run it on the destination host
> under QEMU's x86_64 or i686 emulator. Whether such a  migration is 
> acceptable is totally dependant on the SLA requirements of the application
> running inside the guest.  So you have a simple yes/no answer, but with
> multiple values of 'yes', some better than others.
> 
> In other cases you may not be able to produce a yes/no answer, having to
> apply heuristics. A hueristic may have a firm negative, and a probable
> positive; a probable negative and a firm positive; a probable negative 
> and a probably positive. For example, checking CPU flag compatability.
> If the source has SSE3, and the destinatio nonly have SSE2, you may or
> may not be safely able to migrate depending on whether any app in the
> guest has probed for & is using SSE3 instructions. Most mgmt tools will
> just be conservative in this scenario and refuse to migrate. Or they
> will mask CPU flags to lowest-common denominator.  Another example is
> a guest whose disks are on firewire/usb storage. You can check this and
> if /dev/sda on the source & destination has different model/vendor/serial
> you can say they're different disks. If they are the same model/vendor/
> serial they may or may not be the same physical disks - it is possible
> to get multi-homed firewire disks even if its not common.
> 
> There is also a problem of race conditions in the checking vs action.
> A guest is using 1 GB of ram, and needs to be on a dedicated NUMA node
> with 1 GB ram free. Between the time of performing the check and the 
> guest being migrated the situation may have changed - other guests may
> have auto-ballooned up/down, the kernel itself may have consumed memory
> on the desired NUMA node for its own purposes (disk/io caches), or other
> user apps may have used/released memory. So we can say there's probably
> enough free memory for the guest to migrate and have all its allocations
> on a single node, but we can't easily guarentee it. Do we apply  some
> safety margin in our checks ? eg, check for 1.2 GB free if the guest
> requires 1 GB ? Do we check, and then pre-reserve it, and then check
> again before migrating. Or just accept that some migrations will fail
> and make damn sure the VM is guarentee to keep running safely on its
> original host. Or all of the above
> 
> Finally there is a problem of some compatability factors requiring some
> amount of host 'setup'. If a guest is using iSCSI as its storage, then
> there is a step where the host has to login to the iSCSI target and create
> device nodes for the LUNs before the guest can be run. You don't want 
> every single host to be logged into all your iSCSI targets all the time.
> So what do you do for your migration check wrt to iSCSI ? Do you just
> check that both hosts can access the same iSCSI server + target ? That
> might not detect LUN masking/zoning well enough. So you probably need 
> to actually do all the iSCSI setup on the target before doing the migrate
> compatability check. And if you decide not to migrate after that, you'll
> want to tear the iSCSI stuff down again.
> 
> This all makes it very hard to think of an API for 'checking' migration
> compatability between 2 hosts. The best option I can think of is something
> along the lines of having the application provide a list of 'facts' it
> wants checked, and getting back a list of answers one per fact, with a
> set of values  'no', 'yes', 'probably-yes', 'probably-no', 'no-idea'.
> I'm really not sure it that would even be useful though. Maybe libvirt
> should stick to just providing as rich a set of metadata about all aspects
> of a host & VM as possible, and letting applications do all comparisons.
> Then again I hate the idea of having do duplicate comparisons across all
> apps using libvirt.
> 
> Daniel


Andrew Cathrow
Product Marketing Manager
Red Hat, Inc.

(678) 733 0452 - Mobile
(978) 392-2482 - Office

acathrow at redhat.com
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/libvir-list/attachments/20081028/eb4552e3/attachment-0001.htm>


More information about the libvir-list mailing list