This section describes the installation and configuration of operating systems for the target hosts, as well as deploying SSH keys and configuring storage.
Install one of the following supported operating systems on the target host:
Configure at least one network interface to access the Internet or suitable local repositories.
We recommend adding the Secure Shell (SSH) server packages to the installation on target hosts that do not have local (console) access.
Note
We also recommend setting your locale to en_US.UTF-8. Other locales might work, but they are not tested or supported.
Update package source lists
# apt-get update
Upgrade the system packages and kernel:
# apt-get dist-upgrade
Reboot the host.
Ensure that the kernel version is 3.13.0-34-generic
or later:
# uname -r
Install additional software packages:
# apt-get install bridge-utils debootstrap ifenslave ifenslave-2.6 \
lsof lvm2 chrony openssh-server sudo tcpdump vlan python
Install the kernel extra package if you have one for your kernel version
# apt install linux-image-extra-$(uname -r)
Add the appropriate kernel modules to the /etc/modules
file to
enable VLAN and bond interfaces:
# echo 'bonding' >> /etc/modules
# echo '8021q' >> /etc/modules
Configure Network Time Protocol (NTP) in /etc/chrony/chrony.conf
to
synchronize with a suitable time source and restart the service:
# service chrony restart
Reboot the host to activate the changes and use the new kernel.
Upgrade the system packages and kernel:
# yum upgrade
Disable SELinux. Edit /etc/sysconfig/selinux
, make sure that
SELINUX=enforcing
is changed to SELINUX=disabled
.
Note
SELinux enabled is not currently supported in OpenStack-Ansible for CentOS/RHEL due to a lack of maintainers for the feature.
Reboot the host.
Ensure that the kernel version is 3.10
or later:
# uname -r
Install additional software packages:
# yum install bridge-utils iputils lsof lvm2 \
chrony openssh-server sudo tcpdump python
Add the appropriate kernel modules to the /etc/modules-load.d
file to
enable VLAN and bond interfaces:
# echo 'bonding' >> /etc/modules-load.d/openstack-ansible.conf
# echo '8021q' >> /etc/modules-load.d/openstack-ansible.conf
Configure Network Time Protocol (NTP) in /etc/chrony.conf
to
synchronize with a suitable time source and start the service:
# systemctl enable chronyd.service
# systemctl start chronyd.service
(Optional) Reduce the kernel log level by changing the printk value in your sysctls:
# echo "kernel.printk='4 1 7 4'" >> /etc/sysctl.conf
Reboot the host to activate the changes and use the new kernel.
Upgrade the system packages and kernel:
# zypper up
Reboot the host.
Ensure that the kernel version is 4.4
or later:
# uname -r
Install additional software packages:
# zypper install bridge-utils iputils lsof lvm2 \
chrony opensshr sudo tcpdump python
Add the appropriate kernel modules to the /etc/modules-load.d
file to
enable VLAN and bond interfaces:
# echo 'bonding' >> /etc/modules-load.d/openstack-ansible.conf
# echo '8021q' >> /etc/modules-load.d/openstack-ansible.conf
Configure Network Time Protocol (NTP) in /etc/chrony.conf
to
synchronize with a suitable time source and start the service:
# systemctl enable chronyd.service
# systemctl start chronyd.service
Reboot the host to activate the changes and use the new kernel.
Ansible uses SSH to connect the deployment host and target hosts.
/root/.ssh/authorized_keys
file on each target host.For more information about how to generate an SSH key pair, as well as best practices, see GitHub’s documentation about generating SSH keys.
Important
OpenStack-Ansible deployments require the presence of a
/root/.ssh/id_rsa.pub
file on the deployment host.
The contents of this file is inserted into an
authorized_keys
file for the containers, which is a
necessary step for the Ansible playbooks. You can
override this behavior by setting the
lxc_container_ssh_key
variable to the public key for
the container.
Logical Volume Manager (LVM) enables a single device to be split into multiple logical volumes that appear as a physical storage device to the operating system. The Block Storage (cinder) service, and LXC containers that optionally run the OpenStack infrastructure, can optionally use LVM for their data storage.
Note
OpenStack-Ansible automatically configures LVM on the nodes, and overrides any existing LVM configuration. If you had a customized LVM configuration, edit the generated configuration file as needed.
To use the optional Block Storage (cinder) service, create an LVM
volume group named cinder-volumes
on the storage host. Specify a metadata
size of 2048 when creating the physical volume. For example:
# pvcreate --metadatasize 2048 physical_volume_device_path
# vgcreate cinder-volumes physical_volume_device_path
Optionally, create an LVM volume group named lxc
for container file
systems if you want to use LXC with LVM.
If the lxc
volume group does not exist, containers are
automatically installed on the file system under /var/lib/lxc
by
default.
OpenStack-Ansible uses bridges to connect physical and logical network interfaces on the host to virtual network interfaces within containers. Target hosts need to be configured with the following network bridges:
Bridge name | Best configured on | With a static IP |
---|---|---|
br-mgmt | On every node | Always |
br-storage | On every storage node | When component is deployed on metal |
On every compute node | Always | |
br-vxlan | On every network node | When component is deployed on metal |
On every compute node | Always | |
br-vlan | On every network node | Never |
On every compute node | Never |
For a detailed reference of how the host and container networking is implemented, refer to OpenStack-Ansible Reference Architecture, section Container Networking.
For use case examples, refer to User Guides.
LXC internal: lxcbr0
The lxcbr0
bridge is required for LXC, but OpenStack-Ansible
configures it automatically. It provides external (typically Internet)
connectivity to containers with dnsmasq (DHCP/DNS) + NAT.
This bridge does not directly attach to any physical or logical
interfaces on the host because iptables handles connectivity. It
attaches to eth0
in each container.
The container network that the bridge attaches to is configurable in the
openstack_user_config.yml
file in the provider_networks
dictionary.
Container management: br-mgmt
The br-mgmt
bridge provides management of and
communication between the infrastructure and OpenStack services.
The bridge attaches to a physical or logical interface, typically a
bond0
VLAN subinterface. It also attaches to eth1
in each container.
The container network interface that the bridge attaches to is configurable
in the openstack_user_config.yml
file.
Storage:br-storage
The br-storage
bridge provides segregated access to Block Storage
devices between OpenStack services and Block Storage devices.
The bridge attaches to a physical or logical interface, typically a
bond0
VLAN subinterface. It also attaches to eth2
in each
associated container.
The container network interface that the bridge attaches to is configurable
in the openstack_user_config.yml
file.
OpenStack Networking tunnel: br-vxlan
The br-vxlan
bridge is required if the environment is configured to
allow projects to create virtual networks using VXLAN.
It provides the interface for virtual (VXLAN) tunnel networks.
The bridge attaches to a physical or logical interface, typically a
bond1
VLAN subinterface. It also attaches to eth10
in each
associated container.
The container network interface it attaches to is configurable in
the openstack_user_config.yml
file.
OpenStack Networking provider: br-vlan
The br-vlan
bridge is provides infrastructure for VLAN
tagged or flat (no VLAN tag) networks.
The bridge attaches to a physical or logical interface, typically bond1
.
It attaches to eth11
for VLAN type networks in each associated
container. It is not assigned an IP address because it handles only
layer 2 connectivity.
The container network interface that the bridge attaches to is configurable
in the openstack_user_config.yml
file.
Except where otherwise noted, this document is licensed under Creative Commons Attribution 3.0 License. See all OpenStack Legal Documents.