[libvirt] [RFC PATCH 8/8] qemu: Set cpuset.mems even if the numatune mode is not strict
Osier Yang
jyang at redhat.com
Fri May 10 03:50:33 UTC 2013
On 09/05/13 18:23, Osier Yang wrote:
Missed the address, :\
> [cC Hu Tao]
>
> On 09/05/13 18:22, Osier Yang wrote:
>> When the numatune memory mode is not "strict", the cpuset.mems
>> inherits the parent's setting, which causes problem like:
>>
>> % virsh dumpxml rhel6_local | grep interleave -2
>> <vcpu placement='static'>2</vcpu>
>> <numatune>
>> <memory mode='interleave' nodeset='1-2'/>
>> </numatune>
>> <os>
>>
>> % cat /proc/3713/status | grep Mems_allowed_list
>> Mems_allowed_list: 0-3
>>
>> % virsh numatune rhel6_local
>> numa_mode : interleave
>> numa_nodeset : 0-3
>>
>> Though the domain process's memory binding is set with libnuma
>> after the cgroup setting.
>>
>> The reason for only allowing "strict" mode in current code is the
>> cpuset.mems doesn't understand the memory policy modes (interleave,
>> prefered, strict), it actually equals to the "strict" mode ("strict"
>> means the allocation will fail if the memory cannot be allocated on
>> the target node. Default operation is to fall back to other nodes.
>> From man numa(3)). However, writing the the cpuset.mems even if the
>> numatune memory mode is not strict should be better than the blind
>> inheritance anyway.
>>
>> ---
>> However, I'm not comfortable with the solution, since anyway the
>> modes except "strict" are not meaningful for cpuset.mems.
>>
>> Another problem what I'm not sure about is: If the cpuset.cpus will
>> affect the libnuma setting? Assuming without this patch, domain
>> process's cpuset.mems will be set as '0-7' (8 NUMA nodes, each has 8
>> CPUs). And the numatune memory mode is "interleave", and libnuma set
>> the memory binding as "1-2". Even with this patch applied, setting
>> cpuset.mems as "1-2", any potential problem?
>>
>> So this patch is mainly for raising up the problem, and to see if
>> guys have any opinions. @hutao, since these codes are from you, any
>> opinions/idea? Thanks.
>> ---
>> src/qemu/qemu_cgroup.c | 18 +++++++++++++-----
>> 1 file changed, 13 insertions(+), 5 deletions(-)
>>
>> diff --git a/src/qemu/qemu_cgroup.c b/src/qemu/qemu_cgroup.c
>> index 33eebd7..22fe25b 100644
>> --- a/src/qemu/qemu_cgroup.c
>> +++ b/src/qemu/qemu_cgroup.c
>> @@ -597,11 +597,9 @@ qemuSetupCpusetCgroup(virDomainObjPtr vm,
>> if (!virCgroupHasController(priv->cgroup,
>> VIR_CGROUP_CONTROLLER_CPUSET))
>> return 0;
>> - if ((vm->def->numatune.memory.nodemask ||
>> - (vm->def->numatune.memory.placement_mode ==
>> - VIR_NUMA_TUNE_MEM_PLACEMENT_MODE_AUTO)) &&
>> - vm->def->numatune.memory.mode ==
>> VIR_DOMAIN_NUMATUNE_MEM_STRICT) {
>> -
>> + if (vm->def->numatune.memory.nodemask ||
>> + (vm->def->numatune.memory.placement_mode ==
>> + VIR_NUMA_TUNE_MEM_PLACEMENT_MODE_AUTO)) {
>> if (vm->def->numatune.memory.placement_mode ==
>> VIR_NUMA_TUNE_MEM_PLACEMENT_MODE_AUTO)
>> mem_mask = virBitmapFormat(nodemask);
>> @@ -614,6 +612,16 @@ qemuSetupCpusetCgroup(virDomainObjPtr vm,
>> goto cleanup;
>> }
>> + if (vm->def->numatune.memory.mode ==
>> + VIR_DOMAIN_NUMATUNE_MEM_PREFERRED &&
>> + strlen(mem_mask) != 1) {
>> + virReportError(VIR_ERR_INTERNAL_ERROR, "%s",
>> + _("NUMA memory tuning in 'preferred' mode "
>> + "only supports single node"));
>> + goto cleanup;
>> +
>> + }
>> +
>> rc = virCgroupSetCpusetMems(priv->cgroup, mem_mask);
>> if (rc != 0) {
>
> --
> libvir-list mailing list
> libvir-list at redhat.com
> https://www.redhat.com/mailman/listinfo/libvir-list
More information about the libvir-list
mailing list