Skip to content

A Terraform module to create an Elastic Kubernetes (EKS) cluster and associated worker instances on AWS.

License

Notifications You must be signed in to change notification settings

briskgopesh/terraform-aws-eks

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

terraform-aws-eks

A terraform module to create a managed Kubernetes cluster on AWS EKS. Available through the Terraform registry. Inspired by and adapted from this doc and its source code. Read the AWS docs on EKS to get connected to the k8s dashboard.

Branch Build status
master build Status

Assumptions

  • You want to create an EKS cluster and an autoscaling group of workers for the cluster.
  • You want these resources to exist within security groups that allow communication and coordination. These can be user provided or created within the module.
  • You've created a Virtual Private Cloud (VPC) and subnets where you intend to put the EKS resources.
  • If using the default variable value (true) for configure_kubectl_session, it's required that both kubectl (>=1.10) and aws-iam-authenticator are installed and on your shell's PATH.

Usage example

A full example leveraging other community modules is contained in the examples/eks_test_fixture directory. Here's the gist of using it via the Terraform registry:

module "eks" {
  source                = "terraform-aws-modules/eks/aws"
  cluster_name          = "test-eks-cluster"
  subnets               = ["subnet-abcde012", "subnet-bcde012a"]
  tags                  = {Environment = "test"}
  vpc_id                = "vpc-abcde012"
}

Other documentation

Release schedule

Generally the maintainers will try to release the module once every 2 weeks to keep up with PR additions. If particularly pressing changes are added or maintainers come up with the spare time (hah!), release may happen more often on occasion.

Testing

This module has been packaged with awspec tests through kitchen and kitchen-terraform. To run them:

  1. Install rvm and the ruby version specified in the Gemfile.

  2. Install bundler and the gems from our Gemfile:

    gem install bundler && bundle install
  3. Ensure your AWS environment is configured (i.e. credentials and region) for test.

  4. Test using bundle exec kitchen test from the root of the repo.

For now, connectivity to the kubernetes cluster is not tested but will be in the future. If configure_kubectl_session is set true, once the test fixture has converged, you can query the test cluster from that terminal session with kubectl get nodes --watch --kubeconfig kubeconfig.

Doc generation

Documentation should be modified within main.tf and generated using terraform-docs. Generate them like so:

go get github.com/segmentio/terraform-docs
terraform-docs md ./ | cat -s | tail -r | tail -n +2 | tail -r > README.md

Contributing

Report issues/questions/feature requests on in the issues section.

Full contributing guidelines are covered here.

IAM Permissions

Testing and using this repo requires a minimum set of IAM permissions. Test permissions are listed in the eks_test_fixture README.

Change log

The changelog captures all important release notes.

Authors

Created and maintained by Brandon O'Connor - [email protected]. Many thanks to the contributors listed here!

License

MIT Licensed. See LICENSE for full details.

Inputs

Name Description Type Default Required
cluster_name Name of the EKS cluster. Also used as a prefix in names of related resources. string - yes
cluster_security_group_id If provided, the EKS cluster will be attached to this security group. If not given, a security group will be created with necessary ingres/egress to work with the workers and provide API access to your current IP/32. string `` no
cluster_version Kubernetes version to use for the EKS cluster. string 1.10 no
config_output_path Determines where config files are placed if using configure_kubectl_session and you want config files to land outside the current working directory. Should end in a forward slash / . string ./ no
create_elb_service_linked_role Whether to create the service linked role for the elasticloadbalancing service. Without this EKS cannot create ELBs. string false no
kubeconfig_aws_authenticator_additional_args Any additional arguments to pass to the authenticator such as the role to assume. e.g. ["-r", "MyEksRole"]. list <list> no
kubeconfig_aws_authenticator_command Command to use to to fetch AWS EKS credentials. string aws-iam-authenticator no
kubeconfig_aws_authenticator_env_variables Environment variables that should be used when executing the authenticator. e.g. { AWS_PROFILE = "eks"}. map <map> no
kubeconfig_name Override the default name used for items kubeconfig. string `` no
manage_aws_auth Whether to write and apply the aws-auth configmap file. string true no
map_accounts Additional AWS account numbers to add to the aws-auth configmap. See examples/eks_test_fixture/variables.tf for example format. list <list> no
map_roles Additional IAM roles to add to the aws-auth configmap. See examples/eks_test_fixture/variables.tf for example format. list <list> no
map_users Additional IAM users to add to the aws-auth configmap. See examples/eks_test_fixture/variables.tf for example format. list <list> no
subnets A list of subnets to place the EKS cluster and workers within. list - yes
tags A map of tags to add to all resources. map <map> no
vpc_id VPC where the cluster and workers will be deployed. string - yes
worker_additional_security_group_ids A list of additional security group ids to attach to worker instances list <list> no
worker_group_count The number of maps contained within the worker_groups list. string 1 no
worker_groups A list of maps defining worker group configurations. See workers_group_defaults for valid keys. list <list> no
worker_security_group_id If provided, all workers will be attached to this security group. If not given, a security group will be created with necessary ingres/egress to work with the EKS cluster. string `` no
worker_sg_ingress_from_port Minimum port number from which pods will accept communication. Must be changed to a lower value if some pods in your cluster will expose a port lower than 1025 (e.g. 22, 80, or 443). string 1025 no
workers_group_defaults Override default values for target groups. See workers_group_defaults_defaults in locals.tf for valid keys. map <map> no
write_kubeconfig Whether to write a kubeconfig file containing the cluster configuration. string true no

Outputs

Name Description
cluster_certificate_authority_data Nested attribute containing certificate-authority-data for your cluster. This is the base64 encoded certificate data required to communicate with your cluster.
cluster_endpoint The endpoint for your EKS Kubernetes API.
cluster_id The name/id of the EKS cluster.
cluster_security_group_id Security group ID attached to the EKS cluster.
cluster_version The Kubernetes server version for the EKS cluster.
config_map_aws_auth A kubernetes configuration to authenticate to this EKS cluster.
kubeconfig kubectl config file contents for this EKS cluster.
worker_iam_role_arn default IAM role ARN for EKS worker groups
worker_iam_role_name default IAM role name for EKS worker groups
worker_security_group_id Security group ID attached to the EKS workers.
workers_asg_arns IDs of the autoscaling groups containing workers.
workers_asg_names Names of the autoscaling groups containing workers.

About

A Terraform module to create an Elastic Kubernetes (EKS) cluster and associated worker instances on AWS.

Resources

License

Code of conduct

Stars

Watchers

Forks

Packages

No packages published

Languages

  • HCL 94.7%
  • Smarty 3.1%
  • Ruby 1.3%
  • Shell 0.9%