Installation¶
The networking-odl
repository includes integration with DevStack that
enables creation of a simple OpenDaylight (ODL) development and test
environment. This document discusses what is required for manual installation
and integration into a production OpenStack deployment tool of conventional
architectures that include the following types of nodes:
- Controller - Runs OpenStack control plane services such as REST APIs and databases.
- Network - Provides connectivity between provider (public) and project (private) networks. Services provided include layer-3 (routing), DHCP, and metadata agents. Layer-3 agent is optional. When using netvirt (vpnservice) DHCP/metadata are optional.
- Compute - Runs the hypervisor and layer-2 agent for the Networking service.
ODL Installation¶
http://docs.opendaylight.org provides manual and general documentation for ODL
Review the following documentation regardless of install scenario:
Choose and review one of the following installation scenarios:
- GBP with OpenStack. OpenDaylight Group Based Policy allows users to express network configuration in a declarative rather than imperative way. Often described as asking for “what you want”, rather than “how you can do it”, Group Based Policy achieves this by implementing an Intent System. The Intent System is a process around an intent driven data model and contains no domain specifics but is capable of addressing multiple semantic definitions of intent.
- OVSDB with OpenStack. OpenDaylight OVSDB allows users to take advantage of Network Virtualization using OpenDaylight SDN capabilities whilst utilizing OpenvSwitch. The stack includes a Neutron Northbound, a Network Virtualization layer, an OVSDB southbound plugin, and an OpenFlow southbound plugin.
- VTN with OpenStack. OpenDaylight Virtual Tenant Network (VTN) is an application that provides multi-tenant virtual network on an SDN controller. VTN Manager is implemented as one plugin to the OpenDaylight controller and provides a REST interface to create/update/delete VTN components. It provides an implementation of Openstack L2 Network Functions API.
Networking-odl Installation¶
# sudo pip install networking-odl
Note
pip need to be installed before running above command.
Networking-odl Configuration¶
All related neutron services need to be restarted after configuration change.
Configure Openstack neutron server. The neutron server implements ODL as an ML2 driver. Edit the
/etc/neutron/neutron.conf
file:Enable the ML2 core plug-in.
[DEFAULT] ... core_plugin = neutron.plugins.ml2.plugin.Ml2Plugin
(Optional) Enable ODL L3 router, if QoS feature is desired, then qos should be appended to service_plugins
[DEFAULT] ... service_plugins = odl-router
If the QoS service is enabled then you also need to enable the ODL QoS notification driver.
[qos] ... notification_drivers = odl-qos
..note:
The default notification driver is message_queue, if ODL backend is used, and qos service_plugins and qos extensions drivers are enabled then odl-qos driver should be used to have working QoS functionality.
Configure the ML2 plug-in. Edit the
/etc/neutron/plugins/ml2/ml2_conf.ini
file:Configure the ODL mechanism driver, network type drivers, self-service (tenant) network types, and enable extension drivers(optional).
[ml2] ... mechanism_drivers = opendaylight type_drivers = local,flat,vlan,vxlan tenant_network_types = vxlan extension_drivers = port_security, qos
..note:
The enabling of extension_driver qos is optional, it should be enabled if service_plugins for qos is also enabled.
Configure the vxlan range.
[ml2_type_vxlan] ... vni_ranges = 1:1000
Optionally, enable support for VLAN provider and self-service networks on one or more physical networks. If you specify only the physical network, only administrative (privileged) users can manage VLAN networks. Additionally specifying a VLAN ID range for a physical network enables regular (non-privileged) users to manage VLAN networks. The Networking service allocates the VLAN ID for each self-service network using the VLAN ID range for the physical network.
[ml2_type_vlan] ... network_vlan_ranges = PHYSICAL_NETWORK:MIN_VLAN_ID:MAX_VLAN_ID
Replace
PHYSICAL_NETWORK
with the physical network name and optionally define the minimum and maximum VLAN IDs. Use a comma to separate each physical network.For example, to enable support for administrative VLAN networks on the
physnet1
network and self-service VLAN networks on thephysnet2
network using VLAN IDs 1001 to 2000:network_vlan_ranges = physnet1,physnet2:1001:2000
Enable security groups.
[securitygroup] ... enable_security_group = true
Configure ML2 ODL
[ml2_odl] ... username = <ODL_USERNAME> password = <ODL_PASSWORD> url = http://<ODL_IP_ADDRESS>:<ODL_PORT>/controller/nb/v2/neutron port_binding_controller = pseudo-agentdb-binding
Compute/network nodes¶
Each compute/network node runs the OVS services. If compute/network nodes are already configured to run with Neutron ML2 OVS driver, more steps are necessary. OVSDB with OpenStack can be referred to.
Install the
openvswitch
packages.Start the OVS service.
Using the systemd unit:
# systemctl start openvswitch
Using the
ovs-ctl
script:# /usr/share/openvswitch/scripts/ovs-ctl start
Configure OVS to use ODL as a manager.
# ovs-vsctl set-manager tcp:${ODL_IP_ADDRESS}:6640
Replace
ODL_IP_ADDRESS
with the IP address of ODL controller nodeSet host OVS configurations if port_binding_controller is pseudo-agent
# sudo neutron-odl-ovs-hostconfig
Verify the OVS service.
# ovs-vsctl show
..note:
After setting config files, you have to restart the neutron server if you are using screen then it can be directly started from q-svc window or you can use service neutron-server restart, latter may or may not work depending on OS you are using.