Official AMI Images

RegionVersionInstance TypeArchitectureAMI
ap-northeast-1 v0.8.0 hvmamd64ami-0e282fc6623fc2057
ap-northeast-1 v0.8.0 hvmarm64ami-0d02dc0ebcf8244c0
ap-northeast-2 v0.8.0 hvmarm64ami-061648d404f39d160
ap-northeast-2 v0.8.0 hvmamd64ami-04663739994704f7c
ap-south-1 v0.8.0 hvmamd64ami-08d2c8960ebe55289
ap-south-1 v0.8.0 hvmarm64ami-09fbe45b7ef48d1a6
ap-southeast-1 v0.8.0 hvmamd64ami-0f39d0fe3718d3479
ap-southeast-1 v0.8.0 hvmarm64ami-0a08fa85647f104cd
ap-southeast-2 v0.8.0 hvmamd64ami-06d7823898d83b3c1
ap-southeast-2 v0.8.0 hvmarm64ami-090987b86017ca10c
ca-central-1 v0.8.0 hvmarm64ami-00a9a3cf314429a24
ca-central-1 v0.8.0 hvmamd64ami-01baaa76498316a01
eu-central-1 v0.8.0 hvmamd64ami-0c18141355392bd4f
eu-central-1 v0.8.0 hvmarm64ami-04a86c91ed1ce8860
eu-west-1 v0.8.0 hvmarm64ami-07124251350bf9a99
eu-west-1 v0.8.0 hvmamd64ami-0e1218bc817b4dd27
eu-west-2 v0.8.0 hvmarm64ami-0be2bf0e806d1e342
eu-west-2 v0.8.0 hvmamd64ami-0ada91a91efd9a0b2
eu-west-3 v0.8.0 hvmamd64ami-060cbdf88ec8993c4
eu-west-3 v0.8.0 hvmarm64ami-0d39723609eab1387
sa-east-1 v0.8.0 hvmarm64ami-062a5a013dbc75974
sa-east-1 v0.8.0 hvmamd64ami-02d5b90f8c8408cab
us-east-1 v0.8.0 hvmarm64ami-0b2b789e6f38dabad
us-east-1 v0.8.0 hvmamd64ami-04a401e06c46d717b
us-east-2 v0.8.0 hvmarm64ami-0498315333c6424b9
us-east-2 v0.8.0 hvmamd64ami-0fd143a1c885b37b3
us-west-1 v0.8.0 hvmarm64ami-032ad426fe8a6e1ee
us-west-1 v0.8.0 hvmamd64ami-0e419809080e38f7d
us-west-2 v0.8.0 hvmarm64ami-07e5cfa13ea45de0d
us-west-2 v0.8.0 hvmamd64ami-03fb0b8054d6857d3

Creating a Cluster via the AWS CLI

In this guide we will create an HA Kubernetes cluster with 3 worker nodes. We assume an existing VPC, and some familiarity with AWS. If you need more information on AWS specifics, please see the official AWS documentation.

Create the Subnet

aws ec2 create-subnet \
    --region $REGION \
    --vpc-id $VPC \
    --cidr-block ${CIDR_BLOCK}

Create the AMI

Prepare the Import Prerequisites

Create the S3 Bucket
aws s3api create-bucket \
    --bucket $BUCKET \
    --create-bucket-configuration LocationConstraint=$REGION \
    --acl private
Create the vmimport Role

In order to create an AMI, ensure that the vmimport role exists as described in the official AWS documentation.

Note that the role should be associated with the S3 bucket we created above.

Create the Image Snapshot

First, download the AWS image from a Talos release:

curl -LO | tar -xv

Copy the RAW disk to S3 and import it as a snapshot:

aws s3 cp disk.raw s3://$BUCKET/talos-aws-tutorial.raw
aws ec2 import-snapshot \
    --region $REGION \
    --description "Talos kubernetes tutorial" \
    --disk-container "Format=raw,UserBucket={S3Bucket=$BUCKET,S3Key=talos-aws-tutorial.raw}"

Save the SnapshotId, as we will need it once the import is done. To check on the status of the import, run:

aws ec2 describe-import-snapshot-tasks \
    --region $REGION \

Once the SnapshotTaskDetail.Status indicates completed, we can register the image.

Register the Image
aws ec2 register-image \
    --region $REGION \
    --block-device-mappings "DeviceName=/dev/xvda,VirtualName=talos,Ebs={DeleteOnTermination=true,SnapshotId=$SNAPSHOT,VolumeSize=4,VolumeType=gp2}" \
    --root-device-name /dev/xvda \
    --virtualization-type hvm \
    --architecture x86_64 \
    --ena-support \
    --name talos-aws-tutorial-ami

We now have an AMI we can use to create our cluster. Save the AMI ID, as we will need it when we create EC2 instances.

Create a Security Group

aws ec2 create-security-group \
    --region $REGION \
    --group-name talos-aws-tutorial-sg \
    --description "Security Group for EC2 instances to allow ports required by Talos"

Using the security group ID from above, allow all internal traffic within the same security group:

aws ec2 authorize-security-group-ingress \
    --region $REGION \
    --group-name talos-aws-tutorial-sg \
    --protocol all \
    --port 0 \
    --group-id $SECURITY_GROUP \
    --source-group $SECURITY_GROUP

and expose the Talos and Kubernetes APIs:

aws ec2 authorize-security-group-ingress \
    --region $REGION \
    --group-name talos-aws-tutorial-sg \
    --protocol tcp \
    --port 6443 \
    --cidr \
    --group-id $SECURITY_GROUP
aws ec2 authorize-security-group-ingress \
    --region $REGION \
    --group-name talos-aws-tutorial-sg \
    --protocol tcp \
    --port 50000-50001 \
    --cidr \
    --group-id $SECURITY_GROUP

Create a Load Balancer

aws elbv2 create-load-balancer \
    --region $REGION \
    --name talos-aws-tutorial-lb \
    --type network --subnets $SUBNET

Take note of the DNS name and ARN. We will need these soon.

Create the Machine Configuration Files

Generating Base Configurations

Using the DNS name of the loadbalancer created earlier, generate the base configuration files for the Talos machines:

$ talosctl gen config talos-k8s-aws-tutorial https://<load balancer IP or DNS>:<port>
created init.yaml
created controlplane.yaml
created join.yaml
created talosconfig

At this point, you can modify the generated configs to your liking.

Validate the Configuration Files

$ talosctl validate --config init.yaml --mode cloud
init.yaml is valid for cloud mode
$ talosctl validate --config controlplane.yaml --mode cloud
controlplane.yaml is valid for cloud mode
$ talosctl validate --config join.yaml --mode cloud
join.yaml is valid for cloud mode

Create the EC2 Instances

Note: There is a known issue that prevents Talos from running on T2 instance types. Please use T3 if you need burstable instance types.

Create the Bootstrap Node

aws ec2 run-instances \
    --region $REGION \
    --image-id $AMI \
    --count 1 \
    --instance-type t3.small \
    --user-data file://init.yaml \
    --subnet-id $SUBNET \
    --security-group-ids $SECURITY_GROUP

Create the Remaining Control Plane Nodes

aws ec2 run-instances \
    --region $REGION \
    --image-id $AMI \
    --count 2 \
    --instance-type t3.small \
    --user-data file://controlplane.yaml \
    --subnet-id $SUBNET \
    --security-group-ids $SECURITY_GROUP

Make a note of the resulting PrivateIpAddress from the init and controlplane nodes for later use.

Create the Worker Nodes

aws ec2 run-instances \
    --region $REGION \
    --image-id $AMI \
    --count 3 \
    --instance-type t3.small \
    --user-data file://join.yaml \
    --subnet-id $SUBNET \
    --security-group-ids $SECURITY_GROUP

Configure the Load Balancer

aws elbv2 create-target-group \
    --region $REGION \
    --name talos-aws-tutorial-tg \
    --protocol TCP \
    --port 6443 \
    --vpc-id $VPC

Now, using the target group's ARN, and the PrivateIpAddress from the instances that you created :

aws elbv2 register-targets \
    --region $REGION \
    --target-group-arn $TARGET_GROUP_ARN \
    --targets Id=$CP_NODE_1_IP  Id=$CP_NODE_2_IP  Id=$CP_NODE_3_IP

Using the ARNs of the load balancer and target group from previous steps, create the listener:

aws elbv2 create-listener \
    --region $REGION \
    --load-balancer-arn $LOAD_BALANCER_ARN \
    --protocol TCP \
    --port 443 \
    --default-actions Type=forward,TargetGroupArn=$TARGET_GROUP_ARN

Retrieve the kubeconfig

At this point we can retrieve the admin kubeconfig by running:

talosctl --talosconfig talosconfig config endpoint <control plane 1 IP>
talosctl --talosconfig talosconfig config node <control plane 1 IP>
talosctl --talosconfig talosconfig kubeconfig .