Running the quickstart as an unprivileged user

It is possible to run the bulk of the quickstart deployment as an unprivileged user (a user without root access). In order to do this, there are a few system configuration tasks that must be performed in advance:

  • Making sure required packages are installed
  • Configuring the required libvirt networks

Automatic system configuration

If you want to perform the system configuration tasks manually, skip this section and start reading below at “Configure KVM”.

Place the following into playbook.yml in the tripleo-quickstart directory:

- host: localhost
  roles:
    - environment/setup

And run it like this (assuming that you have sudo access on your local host):

ansible-playbook playbook.yml

Continue reading at Deploying Tripleo.

Configure KVM

You will need to ensure that the kvm kernel module is loaded, and that the appropriate process-specific module (kvm_intel or kvm_amd) is loaded. Run the appropriate modprobe command to load the module:

# modprobe kvm_intel [options...]

Or:

# modprobe kvm_amd [options...]

Where [options...] in the above is either empty, or nested=1 if you want to enable nested kvm.

To ensure this module will be loaded next time your system boots, create /etc/modules-load.d/oooq_kvm.conf with the following content on Intel systems:

kvm_intel

Or on AMD systems:

kvm_amd

If you want to enable nested kvm persistently, create the file /etc/modprobe.d/kvm.conf with the following contents:

options kvm_intel nested=1
options kvm_amd nested=1

Required packages

You will need to install the following packages:

  • qemu-kvm
  • libvirt
  • libvirt-python
  • libguestfs-tools
  • python-lxml

Once these packages are installed, you need to start libvirtd

# systemctl enable libvirtd
# systemctl start libvirtd

Configuring libvirt networks

The quickstart requires two networks. The external network provides inbound access into the virtual environment set up the playbooks. The overcloud network connects the overcloud hosts to the undercloud, and is used both for provisioning, inbound access to the overcloud, and communication between overcloud hosts.

In the following steps, note that the names you choose for the libvirt networks are unimportant (because the vms will be wired up to these networks using bridge names, rather than libvirt network names).

The external network

If you have the standard default libvirt network, you can just use that as your external network. If you would prefer to create a new one, run something like the following:

# virsh net-define /dev/stdin <<EOF
<network>
  <name>external</name>
  <forward mode='nat'>
    <nat>
      <port start='1024' end='65535'/>
    </nat>
  </forward>
  <bridge name='brext' stp='on' delay='0'/>
  <ip address='192.168.23.1' netmask='255.255.255.0'>
    <dhcp>
      <range start='192.168.23.2' end='192.168.23.254'/>
    </dhcp>
  </ip>
</network>
EOF
# virsh net-start external
# virsh net-autostart external

The overcloud network

The overcloud network is really just a bridge, so you could simply configure this through your distributions standard mechanism for configuring persistent bridge devices. You can also do it via libvirt like this:

# virsh net-define /dev/stdin <<EOF
<network>
  <name>overcloud</name>
  <bridge name="brovc" stp='off' delay='0'/>
</network>
EOF
# virsh net-start overcloud
# virsh net-autostart overcloud

Whitelisting bridges

Once you have started the libvirt networks, you need to enter the bridge names in the /etc/qemu/bridge.conf file, which makes these bridges available to unprivileged users via the qemu bridge helper. Note that on some systems this file will be called /etc/qemu-kvm/bridge.conf.

Add an allow line for each bridge you created in the previous steps:

allow brext
allow brovc

Deploying TripleO

With all of the system configuration tasks out of the way, the rest of the process can be run as an unprivileged user. You will need to create a YAML document that described your network configuration and that optionally changes any of the default values used in the quickstart deployment. To describe the network resources we created above, I would create a file called config.yml with the following content:

networks:
  - name: external
    bridge: brext
    address: 192.168.23.1
    netmask: 255.255.255.0

  - name: overcloud
    bridge: brovc

You must have one network named external and one network named overcloud. The address and netmask values must match the values you used to create the libvirt networks.

Place the following into a file playbook.yml in your tripleo-quickstart directory:

- hosts: localhosts
  roles:
    - libvirt/setup
    - tripleo/undercloud
    - tripleo/overcloud

And run it like this:

ansible-playbook playbook.yml -e @config.yml

This will deploy the default virtual infrastructure, which includes an undercloud node, three controllers, one compute node, and one ceph node, and which requires at least 32GB of memory. If you want to deploy a smaller environment, you could use the minimal.yml settings we use in our CI environment:

ansible-playbook playbook.yml -e @config.yml \
  -e playbooks/centosci/minimal.yml

This will create a virtual environment with a single controller and a single compute node, with a total memory footprint of around 22GB.

See Configuration for more information.