Create a GKE Cluster (Workload Identity enabled) using Terraform
In this blog, I will show how Terraform can be used to create a Google Kubernetes Engine (GKE) cluster.

Goal
Create a GKE Cluster which has Workload Identity feature enabled using Terraform.
Prerequisites
This post assumes the following:
1. We already have a GCP Project and a GCS Bucket (we will use this to store Terraform State file) created.
2. Google Kubernetes Engine API is enabled in the GCP Project.
3. Google Cloud SDK (gcloud), kubectl and Terraform is setup on your workstation. If you don’t have, then refer to my previous blogs - Getting started with Terraform and Getting started with Google Cloud SDK.
Source Code
Full code is available on GitHub.
Note: This is not a production-ready codebase.
Create a GKE Cluster
Create a unix directory for the Terraform project.
mkdir ~/terraform-gke-example cd ~/terraform-gke-exampleCreate Terraform configuration file which defines GKE and Google provider.
vi gke.tfThis file has following content
# Specify the GCP Provider provider "google-beta" { project = var.project_id region = var.region version = "~> 3.10" alias = "gb3" } # Create a GKE cluster resource "google_container_cluster" "my_k8s_cluster" { provider = google-beta.gb3 name = "my-k8s-cluster" location = var.region initial_node_count = 1 master_auth { username = "" password = "" } # Enable Workload Identity workload_identity_config { identity_namespace = "${var.project_id}.svc.id.goog" } node_config { machine_type = var.machine_type oauth_scopes = [ "https://www.googleapis.com/auth/logging.write", "https://www.googleapis.com/auth/monitoring", ] metadata = { "disable-legacy-endpoints" = "true" } workload_metadata_config { node_metadata = "GKE_METADATA_SERVER" } labels = { # Update: Replace with desired labels "environment" = "test" "team" = "devops" } } }
Note: workload_identity_config & workload_metadata_config block enables Workload Identity.
Define Terraform variables.
vi variables.tfThis file looks like:
variable "project_id" { description = "Google Project ID." type = string } variable "region" { description = "Google Cloud region" type = string default = "europe-west2" } variable "machine_type" { description = "Google VM Instance type." type = string }
Note: We are defining a default value for region. This means if a value is not supplied for this variable, Terraform will use europe-west2 as its value.
Create a
tfvarsfile.vi terraform.tfvarsproject_id = "" # Put GCP Project ID. machine_type = "n1-standard-1" # Put the desired VM Instance type.Set the remote state.
vi backend.tfterraform { backend "gcs" { bucket = "my-tfstate-bucket" # GCS bucket name to store terraform tfstate prefix = "gke-cluster" # Update to desired prefix name. Prefix name should be unique for each Terraform project having same remote state bucket. } }File structure looks like below
cntekio:~ terraform-gke-example$ tree . ├── backend.tf ├── gke.tf ├── terraform.tfvars └── variables.tf 0 directories, 4 filesNow, initialize the terraform project.
terraform initOutput
Initializing the backend... Successfully configured the backend "gcs"! Terraform will automatically use this backend unless the backend configuration changes. Initializing provider plugins... - Checking for available provider plugins... - Downloading plugin for provider "google-beta" (terraform-providers/google-beta) 3.10.0... Terraform has been successfully initialized! You may now begin working with Terraform. Try running "terraform plan" to see any changes that are required for your infrastructure. All Terraform commands should now work. If you ever set or change modules or backend configuration for Terraform, rerun this command to reinitialize your working directory. If you forget, other commands will detect it and remind you to do so if necessary.
Note: terraform init downloads all the required provider and plugins.
Let’s see the execution plan.
terraform plan --out 1.planOutput
Refreshing Terraform state in-memory prior to plan... The refreshed state will be used to calculate this plan, but will not be persisted to local or remote state storage. ------------------------------------------------------------------------ An execution plan has been generated and is shown below. Resource actions are indicated with the following symbols: + create Terraform will perform the following actions: # google_container_cluster.my_k8s_cluster will be created + resource "google_container_cluster" "my_k8s_cluster" { + additional_zones = (known after apply) + cluster_ipv4_cidr = (known after apply) + default_max_pods_per_node = (known after apply) + enable_binary_authorization = false + enable_intranode_visibility = false + enable_kubernetes_alpha = false + enable_legacy_abac = false + enable_shielded_nodes = false + enable_tpu = false + endpoint = (known after apply) + id = (known after apply) + initial_node_count = 1 + instance_group_urls = (known after apply) + label_fingerprint = (known after apply) + location = "europe-west2" + logging_service = "logging.googleapis.com/kubernetes" + master_version = (known after apply) + monitoring_service = "monitoring.googleapis.com/kubernetes" + name = "my-k8s-cluster" + network = "default" + node_locations = (known after apply) + node_version = (known after apply) + operation = (known after apply) + project = (known after apply) + region = (known after apply) + services_ipv4_cidr = (known after apply) + subnetwork = (known after apply) + tpu_ipv4_cidr_block = (known after apply) + zone = (known after apply) + addons_config { + cloudrun_config { + disabled = (known after apply) } + horizontal_pod_autoscaling { + disabled = (known after apply) } + http_load_balancing { + disabled = (known after apply) } + istio_config { + auth = (known after apply) + disabled = (known after apply) } + kubernetes_dashboard { + disabled = (known after apply) } + network_policy_config { + disabled = (known after apply) } } + authenticator_groups_config { + security_group = (known after apply) } + cluster_autoscaling { + autoscaling_profile = (known after apply) + enabled = (known after apply) + auto_provisioning_defaults { + oauth_scopes = (known after apply) + service_account = (known after apply) } + resource_limits { + maximum = (known after apply) + minimum = (known after apply) + resource_type = (known after apply) } } + database_encryption { + key_name = (known after apply) + state = (known after apply) } + master_auth { + client_certificate = (known after apply) + client_key = (sensitive value) + cluster_ca_certificate = (known after apply) + client_certificate_config { + issue_client_certificate = (known after apply) } } + network_policy { + enabled = (known after apply) + provider = (known after apply) } + node_config { + disk_size_gb = (known after apply) + disk_type = (known after apply) + guest_accelerator = (known after apply) + image_type = (known after apply) + labels = { + "environment" = "test" + "team" = "devops" } + local_ssd_count = (known after apply) + machine_type = "n1-standard-1" + metadata = { + "disable-legacy-endpoints" = "true" } + oauth_scopes = [ + "https://www.googleapis.com/auth/logging.write", + "https://www.googleapis.com/auth/monitoring", ] + preemptible = false + service_account = (known after apply) + taint = (known after apply) + shielded_instance_config { + enable_integrity_monitoring = (known after apply) + enable_secure_boot = (known after apply) } + workload_metadata_config { + node_metadata = "GKE_METADATA_SERVER" } } + node_pool { + initial_node_count = (known after apply) + instance_group_urls = (known after apply) + max_pods_per_node = (known after apply) + name = (known after apply) + name_prefix = (known after apply) + node_count = (known after apply) + node_locations = (known after apply) + version = (known after apply) + autoscaling { + max_node_count = (known after apply) + min_node_count = (known after apply) } + management { + auto_repair = (known after apply) + auto_upgrade = (known after apply) } + node_config { + boot_disk_kms_key = (known after apply) + disk_size_gb = (known after apply) + disk_type = (known after apply) + guest_accelerator = (known after apply) + image_type = (known after apply) + labels = (known after apply) + local_ssd_count = (known after apply) + machine_type = (known after apply) + metadata = (known after apply) + min_cpu_platform = (known after apply) + oauth_scopes = (known after apply) + preemptible = (known after apply) + service_account = (known after apply) + tags = (known after apply) + taint = (known after apply) + sandbox_config { + sandbox_type = (known after apply) } + shielded_instance_config { + enable_integrity_monitoring = (known after apply) + enable_secure_boot = (known after apply) } + workload_metadata_config { + node_metadata = (known after apply) } } + upgrade_settings { + max_surge = (known after apply) + max_unavailable = (known after apply) } } + release_channel { + channel = (known after apply) } + workload_identity_config { + identity_namespace = "workshop-demo-adqw12.svc.id.goog" } } Plan: 1 to add, 0 to change, 0 to destroy. ------------------------------------------------------------------------ This plan was saved to: 1.plan To perform exactly these actions, run the following command to apply: terraform apply "1.plan"
Note: terraform plan output shows that this code is going to a GKE cluster.
--out 1.plan flag tells terraform to save the plan in a file.
The execution plan looks good, so let’s move ahead and apply this plan.
terraform apply 1.planOutput
google_container_cluster.my_k8s_cluster: Creating... google_container_cluster.my_k8s_cluster: Still creating... [10s elapsed] google_container_cluster.my_k8s_cluster: Still creating... [1m10s elapsed] google_container_cluster.my_k8s_cluster: Still creating... [2m10s elapsed] google_container_cluster.my_k8s_cluster: Still creating... [3m20s elapsed] google_container_cluster.my_k8s_cluster: Creation complete after 3m23s [id=projects/workshop-demo-adqw12/locations/europe-west2/clusters/my-k8s-cluster] Apply complete! Resources: 1 added, 0 changed, 0 destroyed.
Note: Creating a GKE cluster could take 4-7mins.
Let’s connect with GKE cluster.
gcloud container clusters get-credentials my-k8s-cluster --region europe-west2 --project workshop-demo-adqw12Output
Fetching cluster endpoint and auth data. kubeconfig entry generated for my-k8s-cluster.Now, run kubectl version command to check server version.
kubectl versionOutput
Client Version: version.Info{Major:"1", Minor:"14+", GitVersion:"v1.14.10-dispatcher", GitCommit:"f5757a1dee5a89cc5e29cd7159076648bf21a02b", GitTreeState:"clean", BuildDate:"2020-02-06T03:31:35Z", GoVersion:"go1.12.12b4", Compiler:"gc", Platform:"darwin/amd64"} Server Version: version.Info{Major:"1", Minor:"14+", GitVersion:"v1.14.10-gke.17", GitCommit:"bdceba0734835c6cb1acbd1c447caf17d8613b44", GitTreeState:"clean", BuildDate:"2020-01-17T23:10:13Z", GoVersion:"go1.12.12b4", Compiler:"gc", Platform:"linux/amd64"}Now, we will move to Google Cloud Platform Console to view GKE cluster properties.

Once we no longer need this infrastructure, we can cleanup to reduce costs.
terraform destroyOutput
google_container_cluster.my_k8s_cluster: Refreshing state... [id=projects/workshop-demo-adqw12/locations/europe-west2/clusters/my-k8s-cluster] An execution plan has been generated and is shown below. Resource actions are indicated with the following symbols: - destroy Terraform will perform the following actions: # google_container_cluster.my_k8s_cluster will be destroyed - resource "google_container_cluster" "my_k8s_cluster" { - additional_zones = [] -> null - cluster_ipv4_cidr = "10.0.0.0/14" -> null - default_max_pods_per_node = 110 -> null - enable_binary_authorization = false -> null - enable_intranode_visibility = false -> null - enable_kubernetes_alpha = false -> null - enable_legacy_abac = false -> null - enable_shielded_nodes = false -> null - enable_tpu = false -> null - endpoint = "34.89.124.93" -> null - id = "projects/workshop-demo-adqw12/locations/europe-west2/clusters/my-k8s-cluster" -> null - initial_node_count = 1 -> null - instance_group_urls = [ - "https://www.googleapis.com/compute/beta/projects/workshop-demo-adqw12/zones/europe-west2-c/instanceGroups/gke-my-k8s-cluster-default-pool-7dc4a362-grp", - "https://www.googleapis.com/compute/beta/projects/workshop-demo-adqw12/zones/europe-west2-b/instanceGroups/gke-my-k8s-cluster-default-pool-0538b086-grp", - "https://www.googleapis.com/compute/beta/projects/workshop-demo-adqw12/zones/europe-west2-a/instanceGroups/gke-my-k8s-cluster-default-pool-858715ad-grp", ] -> null - label_fingerprint = "a9dc16a7" -> null - location = "europe-west2" -> null - logging_service = "logging.googleapis.com/kubernetes" -> null - master_version = "1.14.10-gke.17" -> null - monitoring_service = "monitoring.googleapis.com/kubernetes" -> null - name = "my-k8s-cluster" -> null - network = "projects/workshop-demo-adqw12/global/networks/default" -> null - node_locations = [ - "europe-west2-a", - "europe-west2-b", - "europe-west2-c", ] -> null - node_version = "1.14.10-gke.17" -> null - project = "workshop-demo-adqw12" -> null - resource_labels = {} -> null - services_ipv4_cidr = "10.3.240.0/20" -> null - subnetwork = "projects/workshop-demo-adqw12/regions/europe-west2/subnetworks/default" -> null - addons_config { - network_policy_config { - disabled = true -> null } } - cluster_autoscaling { - autoscaling_profile = "BALANCED" -> null - enabled = false -> null } - database_encryption { - state = "DECRYPTED" -> null } - master_auth { - cluster_ca_certificate = "LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSURERENDQWZTZ0F3SUJBZ0lSQVBmNzJzWRGZhd25uQUYvZz09Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K" -> null - client_certificate_config { - issue_client_certificate = false -> null } } - network_policy { - enabled = false -> null - provider = "PROVIDER_UNSPECIFIED" -> null } - node_config { - disk_size_gb = 100 -> null - disk_type = "pd-standard" -> null - guest_accelerator = [] -> null - image_type = "COS" -> null - labels = { - "environment" = "test" - "team" = "devops" } -> null - local_ssd_count = 0 -> null - machine_type = "n1-standard-1" -> null - metadata = { - "disable-legacy-endpoints" = "true" } -> null - oauth_scopes = [ - "https://www.googleapis.com/auth/logging.write", - "https://www.googleapis.com/auth/monitoring", ] -> null - preemptible = false -> null - service_account = "default" -> null - tags = [] -> null - taint = [] -> null - shielded_instance_config { - enable_integrity_monitoring = true -> null - enable_secure_boot = false -> null } - workload_metadata_config { - node_metadata = "GKE_METADATA_SERVER" -> null } } - node_pool { - initial_node_count = 1 -> null - instance_group_urls = [ - "https://www.googleapis.com/compute/v1/projects/workshop-demo-adqw12/zones/europe-west2-c/instanceGroupManagers/gke-my-k8s-cluster-default-pool-7dc4a362-grp", - "https://www.googleapis.com/compute/v1/projects/workshop-demo-adqw12/zones/europe-west2-b/instanceGroupManagers/gke-my-k8s-cluster-default-pool-0538b086-grp", - "https://www.googleapis.com/compute/v1/projects/workshop-demo-adqw12/zones/europe-west2-a/instanceGroupManagers/gke-my-k8s-cluster-default-pool-858715ad-grp", ] -> null - max_pods_per_node = 0 -> null - name = "default-pool" -> null - node_count = 1 -> null - node_locations = [ - "europe-west2-a", - "europe-west2-b", - "europe-west2-c", ] -> null - version = "1.14.10-gke.17" -> null - management { - auto_repair = false -> null - auto_upgrade = true -> null } - node_config { - disk_size_gb = 100 -> null - disk_type = "pd-standard" -> null - guest_accelerator = [] -> null - image_type = "COS" -> null - labels = { - "environment" = "test" - "team" = "devops" } -> null - local_ssd_count = 0 -> null - machine_type = "n1-standard-1" -> null - metadata = { - "disable-legacy-endpoints" = "true" } -> null - oauth_scopes = [ - "https://www.googleapis.com/auth/logging.write", - "https://www.googleapis.com/auth/monitoring", ] -> null - preemptible = false -> null - service_account = "default" -> null - tags = [] -> null - taint = [] -> null - shielded_instance_config { - enable_integrity_monitoring = true -> null - enable_secure_boot = false -> null } - workload_metadata_config { - node_metadata = "GKE_METADATA_SERVER" -> null } } } - release_channel { - channel = "UNSPECIFIED" -> null } - workload_identity_config { - identity_namespace = "workshop-demo-adqw12.svc.id.goog" -> null } } Plan: 0 to add, 0 to change, 1 to destroy. Do you really want to destroy all resources? Terraform will destroy all your managed infrastructure, as shown above. There is no undo. Only 'yes' will be accepted to confirm. Enter a value: yes google_container_cluster.my_k8s_cluster: Destroying... [id=projects/workshop-demo-adqw12/locations/europe-west2/clusters/my-k8s-cluster] google_container_cluster.my_k8s_cluster: Still destroying... [id=projects/workshop-demo-adqw12/locations/europe-west2/clusters/my-k8s-cluster, 10s elapsed] google_container_cluster.my_k8s_cluster: Still destroying... [id=projects/workshop-demo-adqw12/locations/europe-west2/clusters/my-k8s-cluster, 1m10s elapsed] google_container_cluster.my_k8s_cluster: Still destroying... [id=projects/workshop-demo-adqw12/locations/europe-west2/clusters/my-k8s-cluster, 2m40s elapsed] google_container_cluster.my_k8s_cluster: Destruction complete after 2m41s Destroy complete! Resources: 1 destroyed.
Hope this blog helps you build GKE Cluster Terraform.
If you have a feedback or questions, please reach out to me on LinkedIn or Twitter