Skip to content

Streamline RADAR-base K8s deployment with cloud infrastructure provisioning

License

Notifications You must be signed in to change notification settings

RADAR-base/RADAR-K8s-Infrastructure

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

83 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

RADAR-K8s-Infrastructure

This repository aims to provide IaC templates for RADAR-Kubernetes users who intend to deploy the platform to Kubernetes clusters supported by cloud providers such as AWS.


Terraform validate Terraform validate

Dependencies

Terraform >= 1.7.0, < 1.8.0
AWS CLI >= 2.11

Usage

It is recommended that you use RADAR-K8s-Infrastructure as a template and create your own IaC repository from it (starting with a private one probably). Make sure to customise enclosed templates to your needs before creating the desired infrastructure.

use this template

Configure credentials

export TF_VAR_AWS_REGION=$AWS_REGION
export TF_VAR_AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID
export TF_VAR_AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY
# For temporary credentials and SSO
export TF_VAR_AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN

Workspaces

The definition of resources required for running RADAR-base components is located in the cluster directory, while other optional resources are defined in the config directory. Please treat each directory as a separate workspace and perform terraform operations individually. The cluster resources need to be created and made fully available before you proceed with the creation of the config ones.

To retain the user-specific configurations for future infrastructure updates, modify terraform.tfvars within the workspace and push the change to your repository. If needed, additional variables defined in variables.tf can also be included there.

ℹ️ Important Notice
As a best practice, never save raw values of secret variables in your repository. Instead, always encrypt them before committing. If your cluster is no longer in use, run terraform destory to delete all the associated resources and reduce your cloud spending. If you have resources created within config, run terraform destory in that directory before running the counterpart in cluster.

Create the infrastructure

cd cluster
# Initialise the working directory

terraform init
# Review the changes going to be made 

terraform plan
# Create/update the infrastructure

terraform apply --auto-approve

Created resources:

  • VPC featuring both public and private subnets
  • VPC endpoints for privately accessing AWS services
  • Internet and NAT gateways
  • EKS cluster with a default worker node group
  • EKS coredns, kube-proxy, vpc-cni and aws-ebs-csi-driver addons
  • EBS storage classes referenced by PVCs
  • IRSAs for VPC CNI and EBS CSI controllers
  • Initial EC2 instances launched with Spot capacity
  • Default network ACLs and route tables
  • KMS keys and CloudWatch log groups
  • Essential IAM policies, roles, users and user groups for accessing aforementioned resources

Connect to and verify the cluster

# Make sure to use --region if the cluster is deployed in non-default region and --profile if the cluster is deployed in a non-default AWS account
aws eks update-kubeconfig --name [eks_cluster_name]
kubectl get nodes
kubectl get pods -A

Once the infrastructure update is finished successfully, you can start deploying RADAR-base components to the newly created cluster by following the Installation Guide. Before running helmfile sync, you will find it necessary to configure certain resource values which are required by production.yaml but only known post to infrastructure creation. We have exported the values of those resources and you can get them by simply running:

terraform output

You could also automate this value injection by implementing your own templating strategy to customise production.yaml

Configure the cluster (optional)

N.B.: To get external DNS, Cert Manager and SMTP working via Route 53 (if chosen as your DNS service), you need to configure your registered top-level domain and its corresponding hosted zone ID via variable domain_name in config/terraform.tfvars. Additionally, set enable_route53 to true.

cd config
terraform init
terraform plan
terraform apply --auto-approve

Optional resource creations are disabled by default. To enable the creation of a specific resource named X, navigate to config/terraform.tfvars and update the value of enable_X to true before applying the tempate.

Created resources (if all enabled):

  • EIP allocated for the load balancer created by Ingress-NGINX
  • Karpenter provisioner, the node template and the SQS interruption queue
  • Metrics Server along with the Kubernetes Dashboard and the read-only user
  • MSK cluster featuring Kafka brokers and zookeepers
  • RDS instance running managementportal, appserver and rest_sources_auth databases
  • Route53 zone and records accompanied by IRSAs for external DNS and Cert Manager
  • S3 buckets for intermediate-output-storage, output-storage and velero-backups
  • SES SMTP endpoint
  • CloudWatch event rules and targets
  • Essential IAM policies, roles, users for aforementioned resources

Known limitations

  • Since EBS has been chosen as the default storage, node groups will be created in a single AZ due to the mounting restriction.
  • Sometimes Terraform tries to replace the existing MSK cluster while re-applying the templates even if there is no change on the cluster. Mitigate this with terraform untaint aws_msk_cluster.msk_cluster.
  • Prior to terraform destroy, infrastructure resources created by pods/controllers and may not be visible to Terraform need to be deleted, e.g., nginx-ingress's NLB. A good practice is to always begin by running helmfile destroy.
  • If Karpenter is used for node provisioning, ensure the nodes created by it are not lingering around before running terraform destroy.