[ English | English (United Kingdom) | Deutsch | русский | español | Indonesia ]

Distribution upgrades

This guide provides information about upgrading from one distribution release to the next.

Nota

This guide was last updated when upgrading from Ubuntu Focal to Jammy during the Antelope (2023.1) release. For earlier releases please see other versions of the guide.

Introduction

OpenStack Ansible supports operating system distribution upgrades during specific release cycles. These can be observed by consulting the operating system compatibility matrix, and identifying where two versions of the same operating system are supported.

Upgrades should be performed in the order specified in this guide to minimise the risk of service interruptions. Upgrades must also be carried out by performing a fresh installation of the target system’s operating system, before running openstack-ansible to install services on this host.

Ordering

This guide includes a suggested order for carrying out upgrades. This may need to be adapted dependent on the extent to which you have customised your OpenStack Ansible deployment.

Critically, it is important to consider when you upgrade “repo” hosts/containers. At least one “repo” host should be upgraded before you upgrade any API hosts/containers. The last “repo” host to be upgraded should be the “primary”, and should not be carried out until after the final service which does not support “–limit” is upgraded.

If you have a multi-architecture deployment, then at least one “repo” host of each architecture will need to be upgraded before upgrading any other hosts which use that architecture.

If this order is adapted, it will be necessary to restore some files to the “repo” host from a backup part-way through the process. This will be necessary if no “repo” hosts remain which run the older operating system version, which prevents older packages from being built.

Beyond these requirements, a suggested order for upgrades is a follows:

  1. Infrastructure services (Galera, RabbitMQ, APIs, HAProxy)

    In all cases, secondary or backup instances should be upgraded first

  2. Compute nodes

  3. Network nodes

Pre-Requisites

  • Ensure that all hosts in your target deployment have been installed and configured using a matching version of OpenStack Ansible. Ideally perform a minor upgrade to the latest version of the OpenStack release cycle which you are currently running first in order to reduce the risk of encountering bugs.

  • Check any OpenStack Ansible variables which you customise to ensure that they take into account the new and old operating system version (for example custom package repositories and version pinning).

  • Perform backups of critical data, in particular the Galera database in case of any failures. It is also recommended to back up the “/var/www/repo” directory on the primary “repo” host in case it needs to be restored mid-upgrade.

  • Identify your “primary” HAProxy/Galera/RabbitMQ/repo infrastructure host

    In a simple 3 infrastructure hosts setup, these services/containers usually end up being all on the the same host.

    The “primary” will be the LAST box you’ll want to reinstall.

    • HAProxy/keepalived

      Finding your HAProxy/keepalived primary is as easy as

      ssh {{ external_lb_vip_address }}
      

      Or preferably if you’ve installed HAProxy with stats, like so;

      haproxy_stats_enabled: true
      haproxy_stats_bind_address: "{{ external_lb_vip_address }}"
      

      and can visit https://admin:password@external_lb_vip_address:1936/ and read “Statistics Report for pid # on infrastructure_host”

  • Ensure RabbitMQ is running with all feature flags enabled to avoid conflicts when re-installing nodes. If any are listed as disabled then enable them via the console on one of the nodes:

    rabbitmqctl list_feature_flags
    rabbitmqctl enable_feature_flag all
    

Warnings

  • During the upgrade process, some OpenStack services cannot be deployed by using Ansible’s “–limit”. As such, it will be necessary to deploy some services to mixed operating system versions at the same time.

    The following services are known to lack support for “–limit”:

    • RabbitMQ

    • Repo Server

    • Keystone

  • In the same way as OpenStack Ansible major (and some minor) upgrades, there will be brief interruptions to the entire Galera and RabbitMQ clusters during the upgrade which will result in brief service interruptions.

  • When taking down “memcached” instances for upgrades you may encounter performance issues with the APIs.

Deploying Infrastructure Hosts

  1. Disable HAProxy back ends (optional)

    If you wish to minimise error states in HAProxy, services on hosts which are being reinstalled can be set in maintenance mode (MAINT).

    Log into your primary HAProxy/keepalived and run something similar to

    echo "disable server repo_all-back/<infrahost>_repo_container-<hash>" | socat /var/run/haproxy.stat stdio
    

    for each API or service instance you wish to disable.

    You can also use a playbook from OPS repository like this:

    openstack-ansible set-haproxy-backends-state.yml -e hostname=reinstalled_host -e backend_state=disabled
    

    Or if you’ve enabled haproxy_stats as described above, you can visit https://admin:password@external_lb_vip_address:1936/ and select them and “Set state to MAINT”

  2. Reinstall an infrastructure host’s operating system

    As noted above, this should be carried out for non-primaries first, ideally starting with a “repo” host.

  3. Clearing out stale information

    1. Removing stale ansible-facts

      rm /etc/openstack_deploy/ansible-facts/reinstalled_host*
      

      (* because we’re deleting all container facts for the host as well.)

    2. If RabbitMQ was running on this host

      We forget it by running these commands on another RabbitMQ host.

      rabbitmqctl cluster_status
      rabbitmqctl forget_cluster_node rabbit@removed_host_rabbitmq_container
      
    3. If GlusterFS was running on this host (repo nodes)

      We forget it by running these commands on another repo host. Note that we have to tell Gluster we are intentionally reducing the number of replicas. “N” should be set to the number of repo servers minus 1. Existing gluster peer names can be found using the “gluster peer status” command.

      gluster volume remove-brick gfs-repo replica N removed_host_gluster_peer:/gluster/bricks/1 force
      gluster peer detach removed_host_gluster_peer
      
  4. Do generic preparation of reinstalled host

    openstack-ansible openstack.osa.setup_hosts --limit localhost,reinstalled_host*
    
  5. This step should be executed when you are re-configuring one of haproxy hosts

    Since configuration of haproxy backends happens during individual service provisioning, we need to ensure that all backends are configured before enabling keepalived to select this host.

    Commands below will configure all required backends on haproxy nodes:

    openstack-ansible openstack.osa.haproxy --limit localhost,reinstalled_host --skip-tags keepalived
    openstack-ansible openstack.osa.repo --tags haproxy-service-config
    openstack-ansible openstack.osa.galera_server --tags haproxy-service-config
    openstack-ansible openstack.osa.rabbitmq_server --tags haproxy-service-config
    openstack-ansible openstack.osa.setup_openstack --tags haproxy-service-config
    

    Once this is done, you can deploy keepalived again:

    openstack-ansible openstack.osa.haproxy --tags keepalived --limit localhost,reinstalled_host
    

    After that you might want to ensure that «local» backends remain disabled. You can also use a playbook from OPS repository for this:

    openstack-ansible set-haproxy-backends-state.yml -e hostname=reinstalled_host -e backend_state=disabled --limit reinstalled_host
    
  6. If it is NOT a “primary”, install everything on the new host

    openstack-ansible openstack.osa.setup_infrastructure --limit localhost,repo_all,rabbitmq_all,reinstalled_host*
    openstack-ansible openstack.osa.setup_openstack --limit localhost,keystone_all,reinstalled_host*
    

    (* because we need to include containers in the limit)

  7. If it IS a “primary”, do these steps

    1. Temporarily set your primary Galera in MAINT in HAProxy.

      In order to prevent role from making your primary Galera as UP in haproxy, create an empty file /var/tmp/clustercheck.disabled . You can do this with ad-hoc:

      cd /opt/openstack-ansible
      ansible -m file -a "path=/var/tmp/clustercheck.disabled state=touch" 'reinstalled-host*:&galera_all'
      

      Once it’s done you can run playbook to install MariaDB to the destination

      openstack-ansible openstack.osa.galera_server --limit localhost,reinstalled_host* -e galera_server_bootstrap_node="{{ groups['galera_all'][-1] }}"
      

      You’ll now have mariadb running, and it should be synced with non-primaries.

      To check that verify MariaDB cluster status by executing from host running primary MariaDB following command:

      mysql -e 'SHOW STATUS LIKE "wsrep_cluster_%";'
      

      In case node is not getting synced you might need to restart the mariadb.service and verify everything is in order.

      systemctl restart mariadb.service
      mysql
      mysql> SHOW STATUS LIKE "wsrep_cluster_%";
      mysql> SHOW DATABASES;
      

      Once MariaDB cluster is healthy you can remove the file that disables backend from being used by HAProxy.

      ansible -m file -a "path=/var/tmp/clustercheck.disabled state=absent" 'reinstalled-host_containers:&galera_all'
      
    2. We can move on to RabbitMQ primary

      openstack-ansible openstack.osa.rabbitmq_server -e rabbitmq_primary_cluster_node="{{ hostvars[groups['rabbitmq_all'][-1]]['ansible_facts']['hostname'] }}"
      
    3. Now the repo host primary

      openstack-ansible openstack.osa.repo -e glusterfs_bootstrap_node="{{ groups['repo_all'][-1] }}"
      
    4. Everything should now be in a working state and we can finish it off with

      openstack-ansible openstack.osa.setup_infrastructure --limit localhost,repo_all,rabbitmq_all,reinstalled_host*
      openstack-ansible openstack.osa.setup_openstack --limit localhost,keystone_all,reinstalled_host*
      
  8. Adjust HAProxy status

    If HAProxy was set into MAINT mode, this can now be removed for services which have been restored.

    For the “repo” host, it is important that the freshly installed hosts are set to READY in HAProxy, and any which remain on the old operating system are set to “MAINT”.

    You can also use a playbook from OPS repository to re-enable all backends from the host:

    openstack-ansible set-haproxy-backends-state.yml -e hostname=reinstalled_host -e backend_state=enabled
    

Deploying Compute & Network Hosts

  1. Disable the hypervisor service on compute hosts and migrate any VMs to another available hypervisor.

  2. Reinstall a host’s operating system

  3. Clear out stale ansible-facts

    rm /etc/openstack_deploy/ansible-facts/reinstalled_host*
    

    (* because we’re deleting all container facts for the host as well.)

  4. Execute the following:

    openstack-ansible openstack.osa.setup_hosts --limit localhost,reinstalled_host*
    openstack-ansible openstack.osa.setup_infrastructure --limit localhost,reinstalled_host*
    openstack-ansible openstack.osa.setup_openstack --limit localhost,reinstalled_host*
    

    (* because we need to include containers in the limit)

  5. Re-instate compute node hypervisor UUIDs

    Compute nodes should have their UUID stored in the file “/var/lib/nova/compute_id” and the “nova-compute” service restarted. UUIDs can be found from the command line’openstack hypervisor list”.

    Alternatively, the following Ansible can be used to automate these actions:

    openstack-ansible ../scripts/upgrade-utilities/nova-restore-compute-id.yml --limit reinstalled_host