[Rdo-list] Open vSwitch issues....

Erich Weiler weiler at soe.ucsc.edu
Fri Apr 25 19:11:55 UTC 2014


Actually I appear to have :

openstack-neutron-openvswitch-2014.1-10.el6.noarch

but there appears to be a newer one out there:

openstack-neutron-openvswitch-2014.1-11.el6.noarch.rpm

Is there by chance a bug fix in that one? (assuming this is a bug...)

On 04/25/14 11:50, Erich Weiler wrote:
> Hi Y'all,
>
> I recently began rebuilding my OpenStack installation under the latest
> RDO icehouse release (as of two days ago at least), and everything is
> almost working, but I'm having issues with Open vSwitch, at least on the
> compute nodes.
>
> I'm use the ML2 plugin and VLAN tenant isolation.  I have this in my
> compute node's /etc/neutron/plugin.ini file
>
> ----------
> [ovs]
> bridge_mappings = physnet1:br-eth1
>
> [ml2]
> type_drivers = vlan
> tenant_network_types = vlan
> mechanism_drivers  = openvswitch
>
> [ml2_type_flat]
>
> [ml2_type_vlan]
> network_vlan_ranges = physnet1:200:209
> ----------
>
> My switchports that the nodes connect to are configured as trunks,
> allowing VLANs 200-209 to flow over them.
>
> My network that the VMs should be connecting to is:
>
> # neutron net-show cbse-net
> +---------------------------+--------------------------------------+
> | Field                     | Value                                |
> +---------------------------+--------------------------------------+
> | admin_state_up            | True                                 |
> | id                        | 23028b15-fb12-4a9f-9fba-02f165a52d44 |
> | name                      | cbse-net                             |
> | provider:network_type     | vlan                                 |
> | provider:physical_network | physnet1                             |
> | provider:segmentation_id  | 200                                  |
> | router:external           | False                                |
> | shared                    | False                                |
> | status                    | ACTIVE                               |
> | subnets                   | dd25433a-b21d-475d-91e4-156b00f25047 |
> | tenant_id                 | 7c1980078e044cb08250f628cbe73d29     |
> +---------------------------+--------------------------------------+
>
> # neutron subnet-show dd25433a-b21d-475d-91e4-156b00f25047
> +------------------+--------------------------------------------------+
> | Field            | Value                                            |
> +------------------+--------------------------------------------------+
> | allocation_pools | {"start": "10.200.0.2", "end": "10.200.255.254"} |
> | cidr             | 10.200.0.0/16                                    |
> | dns_nameservers  | 121.43.52.1                                      |
> | enable_dhcp      | True                                             |
> | gateway_ip       | 10.200.0.1                                       |
> | host_routes      |                                                  |
> | id               | dd25433a-b21d-475d-91e4-156b00f25047             |
> | ip_version       | 4                                                |
> | name             |                                                  |
> | network_id       | 23028b15-fb12-4a9f-9fba-02f165a52d44             |
> | tenant_id        | 7c1980078e044cb08250f628cbe73d29                 |
> +------------------+--------------------------------------------------+
>
> So those VMs on that network should send packets that would be tagged
> with VLAN 200.
>
> I launch an instance, then look at the compute node with the instance on
> it.  It doesn't get a DHCP address, so it can't talk to the neutron node
> with the dnsmasq server running on it.  I configure the VM's interface
> to be a static IP on VLAN200, 10.200.0.30, and netmask 255.255.0.0.  I
> have another node set up on VLAN 200 on my switch to test with
> (10.200.0.50) that is a real bare-metal server.
>
> I can't ping my bare-metal server.  I see the packets getting to eth1 on
> my compute node, but stopping there.  Then I figure out that the packets
> are *not being tagged* for VLAN 200 as they leave the compute node!!  So
> the switch is dropping them.    As a test I configure the switchport
> with "native vlan 200", and voila, the ping works.
>
> So, Open vSwitch is not getting that it needs to tag the packets for
> VLAN 200.  A little diagnostics on the compute node:
>
>   ovs-ofctl dump-flows br-int
> NXST_FLOW reply (xid=0x4):
>   cookie=0x0, duration=966.803s, table=0, n_packets=0, n_bytes=0,
> idle_age=966, priority=0 actions=NORMAL
>
> Shouldn't that show some VLAN tagging?
>
> and a tcpdump on eth1 on the compute node:
>
> # tcpdump -e -n -vv -i eth1 | grep -i arp
> tcpdump: WARNING: eth1: no IPv4 address assigned
> tcpdump: listening on eth1, link-type EN10MB (Ethernet), capture size
> 65535 bytes
> 11:21:50.462447 fa:16:3e:94:b3:63 > Broadcast, ethertype ARP (0x0806),
> length 42: Ethernet (len 6), IPv4 (len 4), Request who-has 10.200.0.50
> tell 10.200.0.30, length 28
> 11:21:51.462968 fa:16:3e:94:b3:63 > Broadcast, ethertype ARP (0x0806),
> length 42: Ethernet (len 6), IPv4 (len 4), Request who-has 10.200.0.50
> tell 10.200.0.30, length 28
> 11:21:52.462330 fa:16:3e:94:b3:63 > Broadcast, ethertype ARP (0x0806),
> length 42: Ethernet (len 6), IPv4 (len 4), Request who-has 10.200.0.50
> tell 10.200.0.30, length 28
> 11:21:53.462311 fa:16:3e:94:b3:63 > Broadcast, ethertype ARP (0x0806),
> length 42: Ethernet (len 6), IPv4 (len 4), Request who-has 10.200.0.50
> tell 10.200.0.30, length 28
> 11:21:54.463169 fa:16:3e:94:b3:63 > Broadcast, ethertype ARP (0x0806),
> length 42: Ethernet (len 6), IPv4 (len 4), Request who-has 10.200.0.50
> tell 10.200.0.30, length 28
>
> That tcpdump also confirms the ARP packets are not being tagged 200 as
> they leave the physical interface.
>
> This worked before when I was testing icehouse RC1, I don't know what
> changed with Open vSwitch...  Anyone have any ideas?
>
> Thanks as always for the help!!  This list has been very helpful.
>
> cheers,
> erich




More information about the rdo-list mailing list