Neutron - Networking Service¶
Preparation and deployment¶
Neutron is enabled by default in /etc/kolla/globals.yml
:
#enable_neutron: "{{ enable_openstack_core | bool }}"
Network interfaces¶
Neutron external interface is used for communication with the external world,
for example provider networks, routers and floating IPs.
For setting up the neutron external interface modify
/etc/kolla/globals.yml
setting neutron_external_interface
to the
desired interface name. This interface is used by hosts in the network
group. It is also used by hosts in the compute
group if
enable_neutron_provider_networks
is set or DVR is enabled.
The interface is plugged into a bridge (Open vSwitch or Linux Bridge, depending
on the driver) defined by neutron_bridge_name
, which defaults to br-ex
.
The default Neutron physical network is physnet1
.
Example: single interface¶
In the case where we have only a single Neutron external interface, configuration is simple:
neutron_external_interface: "eth1"
Example: multiple interfaces¶
In some cases it may be necessary to have multiple external network interfaces. This may be achieved via comma-separated lists:
neutron_external_interface: "eth1,eth2"
neutron_bridge_name: "br-ex1,br-ex2"
These two lists are “zipped” together, such that eth1
is plugged into the
br-ex1
bridge, and eth2
is plugged into the br-ex2
bridge. Kolla
Ansible maps these interfaces to Neutron physical networks physnet1
and
physnet2
respectively.
Provider networks¶
Provider networks allow to connect compute instances directly to physical networks avoiding tunnels. This is necessary for example for some performance critical applications. Only administrators of OpenStack can create such networks.
To use provider networks in instances you also need to set the following in
/etc/kolla/globals.yml
:
enable_neutron_provider_networks: yes
For provider networks, compute hosts must have an external bridge
created and configured by Ansible (this is also necessary when
Neutron Distributed Virtual Routing (DVR) mode is enabled). In this case, ensure
neutron_external_interface
is configured correctly for hosts in the
compute
group.
OpenvSwitch (ml2/ovs)¶
By default kolla-ansible
uses openvswitch
as its underlying network
mechanism, you can change that using the neutron_plugin_agent
variable in
/etc/kolla/globals.yml
:
neutron_plugin_agent: "openvswitch"
When using Open vSwitch on a compatible kernel (4.3+ upstream, consult the
documentation of your distribution for support details), you can switch
to using the native OVS firewall driver by employing a configuration override
(see OpenStack Service Configuration in Kolla). You can set it in
/etc/kolla/config/neutron/openvswitch_agent.ini
:
[securitygroup]
firewall_driver = openvswitch
OVN (ml2/ovn)¶
In order to use OVN
as mechanism driver for neutron
, you need to set
the following:
neutron_plugin_agent: "ovn"
When using OVN - Kolla Ansible will not enable distributed floating ip functionality (not enable external bridges on computes) by default. To change this behaviour you need to set the following:
neutron_ovn_distributed_fip: "yes"
Similarly - in order to have Neutron DHCP agents deployed in OVN networking scenario, use:
neutron_ovn_dhcp_agent: "yes"
This might be desired for example when Ironic bare metal nodes are used as a compute service. Currently OVN is not able to answer DHCP queries on port type external, this is where Neutron agent helps.
Mellanox Infiniband (ml2/mlnx)¶
In order to add mlnx_infiniband
to the list of mechanism driver
for neutron
to support Infiniband virtual funtions, you need to
set the following (assuming neutron SR-IOV agent is also enabled using
enable_neutron_sriov
flag):
enable_neutron_mlnx: "yes"
Additionally, you will also need to provide physnet:interface mappings
via neutron_mlnx_physnet_mappings
which is presented to
neutron_mlnx_agent
container via mlnx_agent.ini
and
neutron_eswitchd
container via eswitchd.conf
:
neutron_mlnx_physnet_mappings:
ibphysnet: "ib0"