[ English | Deutsch | español | Indonesia | English (United Kingdom) | русский ]

Jaringan kontainer

OpenStack-Ansible deploys Linux containers (LXC) and uses Linux or Open vSwitch-based bridging between the container and the host interfaces to ensure that all traffic from containers flows over multiple host interfaces. All services in this deployment model use a unique IP address.

This appendix describes how the interfaces are connected and how traffic flows.

For more information about how the OpenStack Networking service (Neutron) uses the interfaces for instance traffic, please see the OpenStack Networking Guide.

Untuk detail tentang konfigurasi jaringan untuk lingkungan Anda, silakan lihat referensi pengaturan openstack_user_config.

Antarmuka host fisik

In a typical production environment, physical network interfaces are combined in bonded pairs for better redundancy and throughput. Avoid using two ports on the same multiport network card for the same bonded interface, because a network card failure affects both of the physical network interfaces used by the bond. Single (bonded) interfaces are also a supported configuration, but will require the use of VLAN subinterfaces.

Linux Bridges/Switches

The combination of containers and flexible deployment options requires implementation of advanced Linux networking features, such as bridges, switches, and namespaces.

  • Bridges provide layer 2 connectivity (similar to physical switches) among physical, logical, and virtual network interfaces within a host. After a bridge/switch is created, the network interfaces are virtually plugged in to it.

    OpenStack-Ansible uses Linux bridges for control plane connections to LXC containers, and can use Linux bridges or Open vSwitch-based bridges for data plane connections that connect virtual machine instances to the physical network infrastructure.

  • Network namespaces provide logically separate layer 3 environments (similar to VRFs) within a host. Namespaces use virtual interfaces to connect with other namespaces, including the host namespace. These interfaces, often called veth pairs, are virtually plugged in between namespaces similar to patch cables connecting physical devices such as switches and routers.

    Setiap kontainer memiliki namespace yang terhubung ke host namespace dengan satu atau lebih pasangan veth ``. Kecuali ditentukan, sistem menghasilkan nama acak untuk pasangan ``veth.

Image berikut menunjukkan bagaimana antarmuka jaringan kontainer terhubung ke jembatan host (host's bridge) dan antarmuka jaringan fisik:

../../_images/networkcomponents.drawio.png

Diagram jaringan

Menghosting dengan layanan yang berjalan dalam kontainer

Diagram berikut menunjukkan bagaimana semua antarmuka dan jembatan saling terhubung untuk menyediakan konektivitas jaringan ke penyebaran OpenStack:

../../_images/networkarch-container-external.drawio.png

The bridge lxcbr0 is configured automatically and provides connectivity for the containers (via eth0) to the outside world, thanks to dnsmasq (dhcp/dns) + NAT.

Catatan

Jika Anda memerlukan konfigurasi jaringan tambahan untuk antarmuka kontainer Anda (seperti mengubah rute pada eth1 untuk rute di jaringan manajemen), silakan sesuaikan file openstack_user_config.yml Anda. Lihat referensi pengaturan openstack_user_config untuk keterangan lebih lanjut.

Lalu lintas Neutron

Common reference drivers, including ML2/LXB, ML2/OVS, and ML2/OVN, and their respective agents, are responsible for managing the virtual networking infrastructure on each node. OpenStack-Ansible refers to Neutron traffic as "data plane" traffic, and can consist of flat, vlan, or overlay technologies such as VXLAN and GENEVE.

Neutron agents can be deployed across a variety of hosts, but are typically limited to dedicated network hosts or infrastructure hosts (controller nodes). Neutron agents are deployed "on metal" and not within an LXC container. Neutron typically requires the operator to define "provider bridge mappings", which map a provider network name to a physical interface. These provider bridge mappings provide flexibility and abstract physical interface names when creating provider networks.

LinuxBridge Example:

bridge_mappings = physnet1:bond1

Open vSwitch/OVN Example:

bridge_mappings = physnet1:br-ex

OpenStack-Ansible provides two overrides when defining provider networks that can be used for creating the mappings and in some cases, connecting the physical interfaces to provider bridges:

  • host_bind_override

  • network_interface

The host_bind_override override is used for LinuxBridge-based deployments, and requires a physical interface name which will then be used by the LinuxBridge agent for flat and vlan-based provider and tenant network traffic.

The network_interface override is used for Open vSwitch and OVN-based deployments, and requires a physical interface name which will be connected to the provider bridge (ie. br-ex) for flat and vlan-based provider and tenant network traffic.

Catatan

Previous versions of OpenStack-Ansible utilized a bridge named br-vlan for flat and vlan-based provider and tenant network traffic. The br-vlan bridge is a leftover of containerized Neutron agents and is no longer useful or recommended.

The following diagrams reflect the differences in the virtual network layout for supported network architectures.

Open Virtual Network (OVN)

Catatan

The ML2/OVN (LXB) mechanism driver is deployed by default as of the Zed release of OpenStack-Ansible.

Networking Node
../../_images/networking-ovn-nn.drawio.png
Compute Node
../../_images/networking-ovn-cn.drawio.png

LinuxBridge

Catatan

The ML2/LinuxBridge (LXB) mechanism driver is marked as "experimental" as of the Zed release of OpenStack-Ansible.

Networking Node
../../_images/networking-linuxbridge-nn.drawio.png
Compute Node
../../_images/networking-linuxbridge-cn.drawio.png

Open vSwitch (OVS)

Networking Node
../../_images/networking-openvswitch-nn.drawio.png
Compute Node
../../_images/networking-openvswitch-cn.drawio.png