[rdo-list] [tripleo] [oooq] Deployment to baremetal fails; "No valid host was found"

Gerard Braad me at gbraad.nl
Mon Jun 20 09:30:13 UTC 2016


Hi,


as mentioned in a previous email, I am deploying baremetal nodes using
the quickstart. At the moment I can introspect nodes correctly, but am
unable to deploy to them.

I performed the checks as mentioned in
/tripleo-docs/doc/source/troubleshooting/troubleshooting-overcloud.rst:


The flavor list I have is unchanged:
[stack at undercloud ~]$ openstack flavor list
+--------------------------------------+---------------+------+------+-----------+-------+-----------+
| ID                                   | Name          |  RAM | Disk |
Ephemeral | VCPUs | Is Public |
+--------------------------------------+---------------+------+------+-----------+-------+-----------+
| 2e72ffb5-c6d7-46fd-ad75-448c0ad6855f | baremetal     | 4096 |   40 |
        0 |     1 | True      |
| 6b8b37e4-618d-4841-b5e3-f556ef27fd4d | oooq_compute  | 8192 |   49 |
        0 |     1 | True      |
| 973b58c3-8730-4b1f-96b2-fda253c15dbc | oooq_control  | 8192 |   49 |
        0 |     1 | True      |
| e22dc516-f53f-4a71-9793-29c614999801 | oooq_ceph     | 8192 |   49 |
        0 |     1 | True      |
| e3dce62a-ac8d-41ba-9f97-84554b247faa | block-storage | 4096 |   40 |
        0 |     1 | True      |
| f5fe9ba6-cf5c-4ef3-adc2-34f3b4381915 | control       | 4096 |   40 |
        0 |     1 | True      |
| fabf81d8-44cb-4c25-8ed0-2afd124425db | compute       | 4096 |   40 |
        0 |     1 | True      |
| fe512696-2294-40cb-9d20-12415f45c1a9 | ceph-storage  | 4096 |   40 |
        0 |     1 | True      |
| ffc859af-dbfd-4e27-99fb-9ab02f4afa79 | swift-storage | 4096 |   40 |
        0 |     1 | True      |
+--------------------------------------+---------------+------+------+-----------+-------+-----------+


In instackenv.json the nodes have been assigned as:
[stack at undercloud ~]$ cat instackenv.json
{
  "nodes":[
  {
    "_comment": "ooo1",
    "pm_type":"pxe_ipmitool",
    "mac": [
        "00:26:9e:9b:c3:36"
    ],
    "cpu": "16",
    "memory": "65536",
    "disk": "370",
    "arch": "x86_64",
    "pm_user":"root",
    "pm_password":"admin",
    "pm_addr":"10.0.108.126",
    "capabilities": "profile:control,boot_option:local"
  },
  {
    "_comment": "ooo2",
    "pm_type":"pxe_ipmitool",
    "mac": [
        "00:26:9e:9c:38:a6"
    ],
    "cpu": "16",
    "memory": "65536",
    "disk": "370",
    "arch": "x86_64",
    "pm_user":"root",
    "pm_password":"admin",
    "pm_addr":"10.0.108.127",
    "capabilities": "profile:compute,boot_option:local"
  }
  ]
}

[stack at undercloud ~]$ ironic node-list
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
| UUID                                 | Name | Instance UUID | Power
State | Provisioning State | Maintenance |
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
| 0956df36-b642-44b8-a67f-0df88270372b | None | None          | power
off   | manageable         | False       |
| cc311355-f373-4e5c-99be-31ba3185639d | None | None          | power
off   | manageable         | False       |
+--------------------------------------+------+---------------+-------------+--------------------+-------------+


And manually I perform the introspection:

[stack at undercloud ~]$ openstack baremetal introspection bulk start
Setting nodes for introspection to manageable...
Starting introspection of node: 0956df36-b642-44b8-a67f-0df88270372b
Starting introspection of node: cc311355-f373-4e5c-99be-31ba3185639d
Waiting for introspection to finish...
Introspection for UUID 0956df36-b642-44b8-a67f-0df88270372b finished
successfully.
Introspection for UUID cc311355-f373-4e5c-99be-31ba3185639d finished
successfully.
Setting manageable nodes to available...
Node 0956df36-b642-44b8-a67f-0df88270372b has been set to available.
Node cc311355-f373-4e5c-99be-31ba3185639d has been set to available.
Introspection completed.


[stack at undercloud ~]$ ironic node-list
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
| UUID                                 | Name | Instance UUID | Power
State | Provisioning State | Maintenance |
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
| 0956df36-b642-44b8-a67f-0df88270372b | None | None          | power
off   | available          | False       |
| cc311355-f373-4e5c-99be-31ba3185639d | None | None          | power
off   | available          | False       |
+--------------------------------------+------+---------------+-------------+--------------------+-------------+


After this, I start the deployment. I have defined the compute and
control flavor to be of the respective type.

[stack at undercloud ~]$ ./overcloud-deploy.sh
<snip>
+ openstack overcloud deploy --templates --timeout 60 --control-scale
1 --control-flavor control --compute-scale 1 --compute-flavor compute
--ntp-server pool.ntp.org -e /tmp/deploy_env.yaml
Deploying templates in the directory /usr/share/openstack-tripleo-heat-templates
2016-06-20 08:18:33 [overcloud]: CREATE_IN_PROGRESS Stack CREATE started
2016-06-20 08:18:33 [HorizonSecret]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:33 [RabbitCookie]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:33 [PcsdPassword]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:33 [MysqlClusterUniquePart]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:33 [MysqlRootPassword]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:33 [Networks]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:34 [VipConfig]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:34 [HeatAuthEncryptionKey]: CREATE_IN_PROGRESS state changed
2016-06-20 08:18:34 [overcloud-VipConfig-i4dgmk37z6hg]:
CREATE_IN_PROGRESS Stack CREATE started
2016-06-20 08:18:34 [overcloud-Networks-4pb3htxq7rkd]:
CREATE_IN_PROGRESS Stack CREATE started
<snip>
2016-06-20 08:19:06 [Controller]: CREATE_FAILED ResourceInError:
resources.Controller: Went to status ERROR due to "Message: No valid
host was found. There are not enough hosts available., Code: 500"
2016-06-20 08:19:06 [Controller]: DELETE_IN_PROGRESS state changed
2016-06-20 08:19:06 [NovaCompute]: CREATE_FAILED ResourceInError:
resources.NovaCompute: Went to status ERROR due to "Message: No valid
host was found. There are not enough hosts available., Code: 500"
2016-06-20 08:19:06 [NovaCompute]: DELETE_IN_PROGRESS state changed
2016-06-20 08:19:09 [Controller]: DELETE_COMPLETE state changed
2016-06-20 08:19:09 [NovaCompute]: DELETE_COMPLETE state changed
2016-06-20 08:19:12 [Controller]: CREATE_IN_PROGRESS state changed
2016-06-20 08:19:12 [NovaCompute]: CREATE_IN_PROGRESS state changed
2016-06-20 08:19:14 [Controller]: CREATE_FAILED ResourceInError:
resources.Controller: Went to status ERROR due to "Message: No valid
host was found. There are not enough hosts available., Code: 500"
2016-06-20 08:19:14 [Controller]: DELETE_IN_PROGRESS state changed
2016-06-20 08:19:14 [NovaCompute]: CREATE_FAILED ResourceInError:
resources.NovaCompute: Went to status ERROR due to "Message: No valid
host was found. There are not enough hosts available., Code: 500"
2016-06-20 08:19:14 [NovaCompute]: DELETE_IN_PROGRESS state changed


But as you can see, the deployment fails.

I check the introspection information and verify that the disk, local
memory and cpus are matching or exceeding the flavor:

[stack at undercloud ~]$ ironic node-show 0956df36-b642-44b8-a67f-0df88270372b
+------------------------+-------------------------------------------------------------------------+
| Property               | Value
                            |
+------------------------+-------------------------------------------------------------------------+
| chassis_uuid           |
                            |
| clean_step             | {}
                            |
| console_enabled        | False
                            |
| created_at             | 2016-06-20T05:51:17+00:00
                            |
| driver                 | pxe_ipmitool
                            |
| driver_info            | {u'ipmi_password': u'******',
u'ipmi_address': u'10.0.108.126',         |
|                        | u'ipmi_username': u'root',
u'deploy_kernel':                            |
|                        | u'07c794a6-b427-4e75-ba58-7c555abbf2f8',
u'deploy_ramdisk': u'67a66b7b- |
|                        | 637f-4b25-bcef-ed39ae32a1f4'}
                            |
| driver_internal_info   | {}
                            |
| extra                  | {u'hardware_swift_object':
u'extra_hardware-0956df36-b642-44b8-a67f-    |
|                        | 0df88270372b'}
                            |
| inspection_finished_at | None
                            |
| inspection_started_at  | None
                            |
| instance_info          | {}
                            |
| instance_uuid          | None
                            |
| last_error             | None
                            |
| maintenance            | False
                            |
| maintenance_reason     | None
                            |
| name                   | None
                            |
| power_state            | power off
                            |
| properties             | {u'memory_mb': u'65536', u'cpu_arch':
u'x86_64', u'local_gb': u'371',   |
|                        | u'cpus': u'16', u'capabilities':
u'profile:control,boot_option:local'}  |
| provision_state        | available
                            |
| provision_updated_at   | 2016-06-20T07:32:46+00:00
                            |
| raid_config            |
                            |
| reservation            | None
                            |
| target_power_state     | None
                            |
| target_provision_state | None
                            |
| target_raid_config     |
                            |
| updated_at             | 2016-06-20T07:32:46+00:00
                            |
| uuid                   | 0956df36-b642-44b8-a67f-0df88270372b
                            |
+------------------------+-------------------------------------------------------------------------+

And also the hypervisor stats are set, but only matching the node count.
[stack at undercloud ~]$ nova hypervisor-stats
+----------------------+-------+
| Property             | Value |
+----------------------+-------+
| count                | 2     |
| current_workload     | 0     |
| disk_available_least | 0     |
| free_disk_gb         | 0     |
| free_ram_mb          | 0     |
| local_gb             | 0     |
| local_gb_used        | 0     |
| memory_mb            | 0     |
| memory_mb_used       | 0     |
| running_vms          | 0     |
| vcpus                | 0     |
| vcpus_used           | 0     |
+----------------------+-------+

Registering the nodes as profile:baremetal has the same effect.

What other parameters are used in making the decision if a node can be
deployed to? I probably miss a small detail... what can I check to
make sure the deployment starts?

regards,


Gerard


-- 

   Gerard Braad | http://gbraad.nl
   [ Doing Open Source Matters ]




More information about the rdo-list mailing list