Logo

Documentation

Deploying Antrea on a GKE cluster

We support running Antrea inside of GKE clusters on Ubuntu Node. Antrea would operate in NetworkPolicy only mode, in which no encapsulation is required for any kind of traffic (Intra Node, Inter Node, etc) and NetworkPolicies are enforced using OVS. Antrea is supported on both VPC-native Enable/Disable modes.

GKE Prerequisites

  1. Install the Google Cloud SDK (gcloud). Refer to Google Cloud SDK installation guide

    curl https://sdk.cloud.google.com | bash
    
  2. Make sure you are authenticated to use the Google Cloud API

    export ADMIN_USER=user@email.com
    gcloud auth login
    
  3. Create a project or use an existing one

    export GKE_PROJECT=gke-clusters
    gcloud projects create $GKE_PROJECT
    

Creating the cluster

You can use any method to create a GKE cluster (gcloud SDK, gcloud Console, etc). The example given here is using the Google Cloud SDK.

Note: Antrea is supported on Ubuntu Nodes only for GKE cluster. Also, it is a must to select service CIDR at the time of cluster deployment.

  1. Create a GKE cluster

    export GKE_ZONE="us-west1"
    export GKE_HOST="UBUNTU"
    export GKE_SERVICE_CIDR="10.94.0.0/16"
    gcloud container --project $GKE_PROJECT clusters create cluster1 --image-type $GKE_HOST \
       --zone $GKE_ZONE --enable-ip-alias --services-ipv4-cidr $GKE_SERVICE_CIDR
    
  2. Access your cluster

    kubectl get nodes
    NAME                                      STATUS   ROLES    AGE     VERSION
    gke-cluster1-default-pool-93d7da1c-61z4   Ready    <none>   3m11s   v1.14.10-gke.17
    gke-cluster1-default-pool-93d7da1c-rkbm   Ready    <none>   3m9s    v1.14.10-gke.17
    
  3. Create a cluster-admin ClusterRoleBinding

    kubectl create clusterrolebinding cluster-admin-binding --clusterrole cluster-admin --user user@email.com
    

    Note: To create clusterRoleBinding, the user must have container.clusterRoleBindings.create permission. Use this command to enable it, if the previous command fails due to permission error. Only cluster Admin can assign this permission.

    gcloud projects add-iam-policy-binding $GKE_PROJECT --member user:user@email.com --role roles/container.admin
    

Deploying Antrea

  1. Prepare the Cluster Nodes

    Deploy antrea-node-init DaemonSet to enable kubelet to operate in CNI mode.

    kubectl apply -f https://raw.githubusercontent.com/vmware-tanzu/antrea/main/build/yamls/antrea-gke-node-init.yml
    
  2. Deploy Antrea

    To deploy a released version of Antrea, pick a deployment manifest from the list of releases. Note that GKE support was added in release 0.5.0, which means you cannot pick a release older than 0.5.0. For any given release <TAG> (e.g. v0.5.0), you can deploy Antrea as follows:

    kubectl apply -f https://github.com/vmware-tanzu/antrea/releases/download/<TAG>/antrea-gke.yml
    

    To deploy the latest version of Antrea (built from the main branch), use the checked-in deployment yaml:

    kubectl apply -f https://raw.githubusercontent.com/vmware-tanzu/antrea/main/build/yamls/antrea-gke.yml
    

    The command will deploy a single replica of Antrea controller to the GKE cluster and deploy Antrea agent to every Node. After a successful deployment you should be able to see these Pods running in your cluster:

    $ kubectl get pods --namespace kube-system  -l app=antrea -o wide
    NAME                                READY   STATUS    RESTARTS   AGE   IP              NODE                                      NOMINATED NODE   READINESS GATES
    antrea-agent-24vwr                  2/2     Running   0          46s   10.138.15.209   gke-cluster1-default-pool-93d7da1c-rkbm   <none>           <none>
    antrea-agent-7dlcp                  2/2     Running   0          46s   10.138.15.206   gke-cluster1-default-pool-9ba12cea-wjzn   <none>           <none>
    antrea-controller-5f9985c59-5crt6   1/1     Running   0          46s   10.138.15.209   gke-cluster1-default-pool-93d7da1c-rkbm   <none>           <none>
    
  3. Restart remaining Pods

    Once Antrea is up and running, restart all Pods in all Namespaces (kube-system, etc) so they can be managed by Antrea.

    $ kubectl delete pods -n kube-system $(kubectl get pods -n kube-system -o custom-columns=NAME:.metadata.name,HOSTNETWORK:.spec.hostNetwork --no-headers=true | grep '<none>' | awk '{ print $1 }')
    pod "event-exporter-v0.2.5-7df89f4b8f-cm5r5" deleted
    pod "fluentd-gcp-scaler-54ccb89d5-2glmv" deleted
    pod "heapster-gke-6dd876579c-fc7xd" deleted
    pod "kube-dns-5877696fb4-7cfbc" deleted
    pod "kube-dns-5877696fb4-9zdpb" deleted
    pod "kube-dns-autoscaler-8687c64fc-h4dtg" deleted
    pod "l7-default-backend-8f479dd9-z42mx" deleted
    pod "metrics-server-v0.3.1-cf56c77fc-7xgvc" deleted
    pod "stackdriver-metadata-agent-cluster-level-6d96ccfd4-5rmwh" deleted
    

Getting Started

To help you get started, see the documentation.