This is the repo used in the Getting into HashiCorp Consul series where we walk through building out a Consul based architecture and cluster, on AWS, from scratch.
This repo is split into branches, each representing a part in the series:
- Part 0 - Beginning of the Project
- Part 1 - Configuring Server and Client on AWS
- Part 2 - Configuring Service Discovery for Consul on AWS
- Part 3 - Scaling, Outage Recovery, and Metrics for Consul on AWS
- Part 4 - Security, Traffice Encryption, and ACLs
- Part 5 - All About Access Control Lists (ACLs)
- Part 6a - Configuring Consul with HCP Vault and Auto-Config
- Part 6b - Mostly Manual Configuration for Part-7 and beyond (use this)
- Part 7 - Enabling Consul Service Mesh
- Part 8 - Traffic Shaping and Envoy Debugging
- Part 9 - Metrics with Prometheus
- Part 10 - Terminating and Ingress Gateways
- Part 11 - Mesh Federation
- Part 12 - Using HCP Consul
- uses HashiCorp Cloud Platform (HCP) Consul instead of deploying Consul via Terraform
- Master - The most up-to-date version of the repo
Note: for Parts 5 and beyond, the corresponding branch will include different resources, setup instructions, and scripts. This is because we build out a foundation with Consul in the earlier parts, but after that, it's been exploration of different aspects of Consul which are not necessarily dependent upon each other. The current master branch is the foundation state where only consul and two services are launched.
To set use this repo, take the following steps:
-
Have an AWS Account.
-
Ensure you have the following things installed locally:
-
Either use the root user for your account, or create a new IAM user with either Admin or PowerUser permissions.
-
Set up AWS credentials locally either through environment variables, through the AWS CLI, or directly in
~/.aws/credentials
and~/.aws/config
. More information on authenticating with AWS for Terraform. -
Create an EC2 Keypair, download the key, and add the private key identity to the auth agent. More information on creating an EC2 Keypair.
# After downloading the key from AWS, on Mac for example chmod 400 ~/Downloads/your_aws_ec2_key.pem # Optionally move it to another directory mv ~/Downloads/your_aws_ec2_key.pem ~/.ssh/ # Add the key to your auth agent ssh-add -k ~/.ssh/your_aws_ec2_key.pem
-
Create a
terraform.tfvars
file and add the name of your key for theec2_key_pair_name
variable:ec2_key_pair_name = "your_aws_ec2_key"
-
Run
terraform apply
! -
After the apply is complete, run the post apply script:
# this will output two things: # 1. Sensitive values needed in a local file 'tokens.txt' # 2. Values required by the metrics_module # 3. Detailed setup instructions which are also listed below bash scripts/post-apply.sh
-
SSH into
bastion
and then into yourgetting-into-consul-api
nodes...- Add the
client_api_node_id_token
fromtokens.txt
to the/etc/consul.d/consul.hcl
file in the acl.tokens block. - Add the
client_api_service_token
fromtokens.txt
to the/etc/consul.d/api.hcl
file in the service.token block. - Add the
client_api_service_token
fromtokens.txt
to the/etc/systemd/system/consul-envoy.service
. - Restart both
consul
,api
, andconsul-envoy
service:sudo systemctl daemon-reload sudo systemctl restart consul api consul-envoy
NOTE: Sometimes
consul-envoy
will fail to start ifconsul
isn't given enough time to start up. Simply restartconsul-envoy
again if this is the case. - Add the
-
SSH into
bastion
and then into yourgetting-into-consul-web
nodes...- Add the
client_web_node_id_token
fromtokens.txt
to the/etc/consul.d/consul.hcl
file in the acl.tokens block. - Add the
client_web_service_token
fromtokens.txt
to the/etc/consul.d/web.hcl
file in the service.token block. - Add the
client_web_service_token
fromtokens.txt
to the/etc/systemd/system/consul-envoy.service
. - Restart both
consul
,web
, andconsul-envoy
service:sudo systemctl daemon-reload sudo systemctl restart consul web consul-envoy
NOTE: Sometimes
consul-envoy
will fail to start ifconsul
isn't given enough time to start up. Simply restartconsul-envoy
again if this is the case. - Add the
-
Head to the Consul UI via your
consul_server
output from Terraform (theapplication load balancer
DNS for the server).- Login with your root token (the
consul_token
output, you can find it in your state file)
- Login with your root token (the
-
To verify everything is working, check out your Consul UI...
- All services in the Services tab should be green.
- All nodes in the Nodes tab should be green.
-
To verify the web service is up and running, head to the DNS printed in the terraform output as
web_server
- It should show the upstream
body
of theapi
server with an IP address indc1
.
- It should show the upstream
There's also another module nested in this repository that will stand up a prometheus deployment to monitor your Consul cluster. To deploy it.
-
Ensure that the above "Getting Started" instructions have been followed.
- including running the
post-apply.sh
script that creates needed variables for the metrics deployment.
- including running the
-
Run the
post-apply-metrics.sh
script:bash scripts/post-apply-metrics.sh
-
Navigate to the nested
metrics_module
.cd metrics_module/
-
Initialize the nested Terraform Module:
terraform init
-
Deploy it:
terraform apply
-
Afterwards, it'll output
metrics_endpoint
which is the endpoint you can visit to view your metrics.
Although this repo is set up so that you can get everything working via terraform apply
, if you'd like to take the manual steps for learning, you can reference these documents:
- From Part 1 to Part 2 Manual Steps
- From Part 2 to Part 3 Manual Steps
- From Part 3 to Part 4 Manual Steps
- From Part 4 to Part 5 Manual Steps
- Follow the Steps on this README to get to Part 6
- From Part 6 to Part 7 Manual Steps
- From Part 7 to Part 8 Manual Steps
- From Part 8 to Part 9 Manual Steps
- Checkout the part-9-manual branch to follow these.
- From Part 9 to Part 10 Manual Steps
- From Part 11 to Part 12 Manual Steps
For example, if you wanted to manually learn Part 1 to Part 2, begin on the Part 1 Branch, and follow the "From Part 1 to Part 2 Manual Steps".
The most likely cause of this is a failure to fetch and install consul on the servers due to a failure to get the required GPG key. The most straightforward way to fix this is to terraform destroy
the infrastructure and reapply it via terraform apply
.
If everything deployed fine and you can see the Consul UI and the web
service is reachable but is saying that the api
can't be reached, it's likely Consul Intentions. To fix:
- Login with your consul bootstrap token.
- Click the Web service and then click on the Topology tab.
- Click on the red arrow between the web and the api boxes and click Create to create a new intention that allows the
web
to access theapi
service.
- Cloud Auto-Join is set up for part 1, despite not being in the stream itself.