- Getting Started
- Support for K8s Installers
- Deploying on Kind
- Deploying on Minikube
- Installing with Helm
- Antrea Network Policy
- Traffic Encryption (Ipsec / WireGuard)
- Securing Control Plane
- Security considerations
- OS-specific Known Issues
- OVS Pipeline
- Feature Gates
- Antrea Proxy
- Network Flow Visibility
- Traceflow Guide
- NoEncap and Hybrid Traffic Modes
- Egress Guide
- NodePortLocal Guide
- Antrea IPAM Guide
- Exposing Services of type LoadBalancer
- Traffic Control
- Antrea API Groups
- Antrea API Reference
Antrea is super easy to install. All the Antrea components are containerized and can be installed using the Kubernetes deployment manifest.
Ensuring requirements are satisfied
Antrea relies on
NodeIPAM for per-Node CIDR allocation.
NodeIPAM can run
within the Kubernetes
kube-controller-manager, or within the Antrea
NodeIPAM within kube-controller-manager
kubeadm to create the Kubernetes cluster, passing
--pod-network-cidr=<CIDR Range for Pods> to
kubeadm init will enable
NodeIpamController. Clusters created with kubeadm will always have
CNI plugins enabled. Refer to
Creating a cluster with kubeadm
for more information about setting up a Kubernetes cluster with
When the cluster is deployed by other means then:
kube-controller-managershould be started with the following flags:
--cluster-cidr=<CIDR Range for Pods>
kubeletshould be started with the
To enable masquerading of traffic for Service cluster IP via iptables,
kube-proxyshould be started with the
--cluster-cidr=<CIDR Range for Pods>flag.
NodeIPAM within Antrea Controller
For further info about running NodeIPAM within Antrea Controller, see Antrea IPAM Capabilities
As for OVS, when using the built-in kernel module, kernel version >= 4.6 is required. On the other hand, when building it from OVS sources, OVS version >= 2.6.0 is required.
Red Hat Enterprise Linux and CentOS 7.x use kernel 3.10, but as changes to OVS kernel modules are regularly backported to these kernel versions, they should work with Antrea, starting with version 7.4.
In case a node does not have a supported OVS module installed,
you can install it following the instructions at:
Installing Open vSwitch.
Please be aware that the
vport-stt module is not in the Linux tree and needs to be
built from source, please build and load it manually before STT tunneling is enabled.
Some experimental features disabled by default may have additional requirements, please refer to the Feature Gates documentation to determine whether it applies to you.
Antrea will work out-of-the-box on most popular Operating Systems. Known issues encountered when running Antrea on specific OSes are documented here.
There are also a few network prerequisites which need to be satisfied, and they depend on the tunnel mode you choose, please check network requirements.
To deploy a released version of Antrea, pick a deployment manifest from the
list of releases. For any
v0.1.0), you can deploy Antrea as follows:
kubectl apply -f https://github.com/antrea-io/antrea/releases/download/<TAG>/antrea.yml
To deploy the latest version of Antrea (built from the main branch), use the checked-in deployment yaml:
kubectl apply -f https://raw.githubusercontent.com/antrea-io/antrea/main/build/yamls/antrea.yml
Antrea supports some experimental features that can be enabled or disabled, please refer to the Feature Gates documentation for more information.
If you want to add Windows Nodes to your cluster, please refer to these installation instructions.
Starting with v1.0, Antrea supports arm64 and arm/v7 Nodes. The installation
instructions do not change when some (or all) Linux Nodes in a cluster use an
ARM architecture: the same deployment YAML can be used, as the
antrea/antrea-ubuntu Docker image is actually a manifest list with support for
the amd64, arm64 and arm/v7 architectures.
Note that while we do run a subset of the Kubernetes conformance tests on both the arm/v7 and arm64 Docker images (using k3s as the Kubernetes distribution), our testing is not as thorough as for the amd64 image. However, we do not anticipate any issue.
Install with Helm
Starting with v1.8, Antrea can be installed and updated with Helm. Please refer to these installation instructions.
Deploying Antrea on a Cluster with Existing CNI
The instructions above only apply when deploying Antrea in a new cluster. If you need to migrate your existing cluster from another CNI plugin to Antrea, you will need to do the following:
- Delete previous CNI, including all resources (K8s objects, iptables rules, interfaces, …) created by that CNI.
- Deploy Antrea.
- Restart all Pods in the CNI network in order for Antrea to set-up networking
for them. This does not apply to Pods which use the Node’s network namespace
(i.e. Pods configured with
hostNetwork: true). You may use
kubectl drainto drain each Node or reboot all your Nodes.
While this is in-progress, networking will be disrupted in your cluster. After deleting the previous CNI, existing Pods may not be reachable anymore.
For example, when migrating from Flannel to Antrea, you will need to do the following:
- Delete Flannel with
kubectl delete -f <path to your Flannel YAML manifest>.
- Delete Flannel bridge and tunnel interface with
ip link delete flannel.1 && ip link delete flannel cni0on each Node.
- Ensure requirements are satisfied.
- Deploy Antrea.
- Drain and uncordon Nodes one-by-one. For each Node, run
kubectl drain --ignore-daemonsets <node name> && kubectl uncordon <node name>. The
--ignore-daemonsetsflag will ignore DaemonSet-managed Pods, including the Antrea Agent Pods. If you have any other DaemonSet-managed Pods (besides the Antrea ones and system ones such as kube-proxy), they will be ignored and will not be drained from the Node. Refer to the Kubernetes documentation for more information. Alternatively, you can also restart all the Pods yourself, or simply reboot your Nodes.
To build the image locally, you can follow the instructions in the Contributor Guide.
Deploying Antrea in Kind
Deploying Antrea in Minikube
Deploying Antrea in AKS, EKS, and GKE
Antrea can work with cloud managed Kubernetes services, and can be deployed to AKS, EKS, and GKE clusters.
- To deploy Antrea to an AKS or an AKS Engine cluster, please refer to the AKS installation guide.
- To deploy Antrea to an EKS cluster, please refer to the EKS installation guide.
- To deploy Antrea to a GKE cluster, please refer to the GKE installation guide.
Deploying Antrea with Custom Certificates
By default, Antrea generates the certificates needed for itself to run. To provide your own certificates, please refer to Securing Control Plane.
Antctl: Installation and Usage
To use antctl, the Antrea command-line tool, please refer to this guide.
Antrea Network Policy
Besides Kubernetes NetworkPolicy, Antrea also implements its own Network Policy CRDs, which provide advanced features including: policy priority, tiering, deny action, external entity, and policy statistics. For more information on usage of Antrea Network Policies, refer to the Antrea Network Policy document.
Antrea supports specifying which egress (SNAT) IP the traffic from the selected Pods to the external network should use and which Node the traffic should leave the cluster from. For more information, refer to the Egress document.
Network Flow Visibility
Antrea supports exporting network flow information using IPFIX, and provides a reference cookbook on how to visualize the exported network flows using Elastic Stack and Kibana dashboards. For more information, refer to the network flow visibility document.
NoEncap and Hybrid Traffic Modes
Besides the default
Encap mode, in which Pod traffic across Nodes will be
encapsulated and sent over tunnels, Antrea also supports
traffic modes. In
NoEncap mode, Antrea does not encapsulate Pod traffic, but
relies on the Node network to route the traffic across Nodes. In
Antrea encapsulates Pod traffic when the source Node and the destination Node
are in different subnets, but does not encapsulate when the source and the
destination Nodes are in the same subnet. Refer to
to learn how to configure Antrea with
Antrea Web UI
Antrea comes with a web UI, which can show runtime information of Antrea components and perform Antrea Traceflow operations. Please refer to the Antrea UI repository for installation instructions and more information.
OVS Hardware Offload
Antrea can offload OVS flow processing to the NICs that support OVS kernel hardware offload using TC. The hardware offload can improve OVS performance significantly. For more information on how to configure OVS offload, refer to the OVS hardware offload guide.
Antrea supports exporting metrics to Prometheus. For more information, refer to the Prometheus integration document.
Support for Services of type LoadBalancer
By leveraging Antrea’s Service external IP management feature or configuring MetalLB to work with Antrea, Services of type LoadBalancer can be supported without requiring an external LoadBalancer. To learn more information, please refer to the Service LoadBalancer document.
Traceflow is a very useful network diagnosis feature in Antrea. It can trace and report the forwarding path of a specified packet in the Antrea network. For usage of this feature, refer to the Traceflow user guide.
Antrea supports encrypting traffic between Linux Nodes using IPsec or WireGuard. To deploy Antrea with traffic encryption enabled, please refer to this guide.
Antrea Multi-cluster implements Multi-cluster Service API, which allows users to create multi-cluster Services that can be accessed cross clusters in a ClusterSet. Antrea Multi-cluster also supports Antrea ClusterNetworkPolicy replication. Multi-cluster admins can define ClusterNetworkPolicies to be replicated across the entire ClusterSet, and enforced in all member clusters. To learn more information about Antrea Multi-cluster, please refer to the Antrea Multi-cluster user guide.