Documentation
Introduction
Cloud Deployment
Reference
- Antrea Network Policy
- antctl
- Architecture
- IPsec Configuration
- Securing Control Plane
- Troubleshooting
- OS-specific Known Issues
- OVS Pipeline
- Feature Gates
- Network Flow Visibility
- Traceflow Guide
- NoEncap and Hybrid Traffic Modes
- Egress Guide
- NodePortLocal Guide
- Versioning
- Antrea API Groups
- Antrea API Reference
Windows
Integrations
Cookbooks
Developer Guide
Project Information
Deploying Antrea on a GKE cluster
We support running Antrea inside of GKE clusters on Ubuntu Node. Antrea would operate in NetworkPolicy only mode, in which no encapsulation is required for any kind of traffic (Intra Node, Inter Node, etc) and NetworkPolicies are enforced using OVS. Antrea is supported on both VPC-native Enable/Disable modes.
GKE Prerequisites
-
Install the Google Cloud SDK (gcloud). Refer to Google Cloud SDK installation guide
curl https://sdk.cloud.google.com | bash
-
Make sure you are authenticated to use the Google Cloud API
export ADMIN_USER=user@email.com gcloud auth login
-
Create a project or use an existing one
export GKE_PROJECT=gke-clusters gcloud projects create $GKE_PROJECT
Creating the cluster
You can use any method to create a GKE cluster (gcloud SDK, gcloud Console, etc). The example given here is using the Google Cloud SDK.
Note: Antrea is supported on Ubuntu Nodes only for GKE cluster. Also, it is a must to select service CIDR at the time of cluster deployment.
-
Create a GKE cluster
export GKE_ZONE="us-west1" export GKE_HOST="UBUNTU" export GKE_SERVICE_CIDR="10.94.0.0/16" gcloud container --project $GKE_PROJECT clusters create cluster1 --image-type $GKE_HOST \ --zone $GKE_ZONE --enable-ip-alias --services-ipv4-cidr $GKE_SERVICE_CIDR
-
Access your cluster
kubectl get nodes NAME STATUS ROLES AGE VERSION gke-cluster1-default-pool-93d7da1c-61z4 Ready <none> 3m11s v1.14.10-gke.17 gke-cluster1-default-pool-93d7da1c-rkbm Ready <none> 3m9s v1.14.10-gke.17
-
Create a cluster-admin ClusterRoleBinding
kubectl create clusterrolebinding cluster-admin-binding --clusterrole cluster-admin --user user@email.com
Note: To create clusterRoleBinding, the user must have
container.clusterRoleBindings.create
permission. Use this command to enable it, if the previous command fails due to permission error. Only cluster Admin can assign this permission.gcloud projects add-iam-policy-binding $GKE_PROJECT --member user:user@email.com --role roles/container.admin
Deploying Antrea
-
Prepare the Cluster Nodes
Deploy
antrea-node-init
DaemonSet to enablekubelet
to operate in CNI mode.kubectl apply -f https://raw.githubusercontent.com/antrea-io/antrea/main/build/yamls/antrea-gke-node-init.yml
-
Deploy Antrea
To deploy a released version of Antrea, pick a deployment manifest from the list of releases. Note that GKE support was added in release 0.5.0, which means you cannot pick a release older than 0.5.0. For any given release
<TAG>
(e.g.v0.5.0
), you can deploy Antrea as follows:kubectl apply -f https://github.com/antrea-io/antrea/releases/download/<TAG>/antrea-gke.yml
To deploy the latest version of Antrea (built from the main branch), use the checked-in deployment yaml:
kubectl apply -f https://raw.githubusercontent.com/antrea-io/antrea/main/build/yamls/antrea-gke.yml
The command will deploy a single replica of Antrea controller to the GKE cluster and deploy Antrea agent to every Node. After a successful deployment you should be able to see these Pods running in your cluster:
$ kubectl get pods --namespace kube-system -l app=antrea -o wide NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES antrea-agent-24vwr 2/2 Running 0 46s 10.138.15.209 gke-cluster1-default-pool-93d7da1c-rkbm <none> <none> antrea-agent-7dlcp 2/2 Running 0 46s 10.138.15.206 gke-cluster1-default-pool-9ba12cea-wjzn <none> <none> antrea-controller-5f9985c59-5crt6 1/1 Running 0 46s 10.138.15.209 gke-cluster1-default-pool-93d7da1c-rkbm <none> <none>
-
Restart remaining Pods
Once Antrea is up and running, restart all Pods in all Namespaces (kube-system, etc) so they can be managed by Antrea.
$ kubectl delete pods -n kube-system $(kubectl get pods -n kube-system -o custom-columns=NAME:.metadata.name,HOSTNETWORK:.spec.hostNetwork --no-headers=true | grep '<none>' | awk '{ print $1 }') pod "event-exporter-v0.2.5-7df89f4b8f-cm5r5" deleted pod "fluentd-gcp-scaler-54ccb89d5-2glmv" deleted pod "heapster-gke-6dd876579c-fc7xd" deleted pod "kube-dns-5877696fb4-7cfbc" deleted pod "kube-dns-5877696fb4-9zdpb" deleted pod "kube-dns-autoscaler-8687c64fc-h4dtg" deleted pod "l7-default-backend-8f479dd9-z42mx" deleted pod "metrics-server-v0.3.1-cf56c77fc-7xgvc" deleted pod "stackdriver-metadata-agent-cluster-level-6d96ccfd4-5rmwh" deleted