This repository contains the Terraform configuration for the infrastructure of the Private Rented Sector Database (PRSDB) service. The main respository for the service, which includes Architecture Decision Records for this infrastructure can be found at https://github.com/communitiesuk/prsdb-webapp
Install the latest AWS vault by following these instructions. (N.B. there is no windows AMD64 version, so you need to use 386 in that case). You will need to rename the .exe file to "aws-vault.exe" and add it to your PATH.
Install AWS CLI by following these instructions.
Install the Session Manager plugin for AWS CLI by following these instructions You may need to add the "session-manager-plugin.exe" file to your PATH.
Before setting up access via SSO, delete any existing mhclg elements from your ~/.aws/config
file (or delete the whole file if you're only using it for this project) to prevent any potential aws-vault authentication issues.
Run aws configure sso
and follow the prompts:
SSO session name (Recommended): mhclg-sso
SSO start URL [None]: https://d-9c67685a87.awsapps.com/start
SSO region [None]: eu-west-2
SSO registration scopes: sso:account:access
The AWS CLI will provide you with a code in your terminal and open a browser window. (If you're asked to) log in using your Super User MHCLG account then confirm the code in the browser matches. On the subsequent page you should click 'Allow access' then return to the terminal.
Note: To prevent authentication errors, use a browser that is not already logged into AWS via SSO (e.g. use a different profile in Microsoft Edge).
The terminal should now display the list of AWS accounts you have access to. Select the integration account. As only the developer role is available, you should then be prompted to provide the following information:
CLI default client [None]: eu-west-2
CLI default output format: can be left blank - press enter
CLI profile name [name_of_role-account_id]: mhclg-int
In your terminal, navigate to the integration directory and run aws-vault exec mhclg-int -- <your preferred shell>
.
This will start a sub-shell with the access key credentials available and the role of that profile. Use exit
to return to your previous shell when you are done.
If you want to work on a different environment, repeat this process with a different AWS account and profile name.
When using SSO, you must specify the profile name in the command. For example, aws s3 ls --profile <your profile>
.
You can end your session with aws sso logout
and log back in with aws sso login --profile <your profile>
.
Use the shell script "setup/create_profiles.sh" (or "setup/create_profiles.ps1" if using Windows Powershell) to add a .aws/config
file to your home directory, with your mfa identifier value added.
You can find this in the aws console under the top right drop down menu -> Security credentials
.
Note you may need to switch out of any roles you are switched into to see the Security credentials
option
Your mfa identifier look like arn:aws:iam::123456789012:mfa/yourname
.
You should not override the destination file.
setup/create_profiles.sh "Your mfa_serial value" ["Override destination file"]
(Powershell)
setup/create_profiles.ps1 "Your mfa_serial value" ["Override destination file"]
(If you get a "Could not find a part of the path..." error when the script runs, you may need to create the .aws folder in your home directory first)
Create an access key in the AWS portal on the Security credentials page (just below where you found you mfa_serial
). In your terminal, run aws-vault add mhclg
and provide your access key details.
You should now be able to start an aws-vault exec
session using the instructions from the section above.
Install the appropriate version of Terraform by following these instructions. We are currently using 1.9.x, so make sure you select that version. If using the "Manual installation" method, you can select the correct version from the dropdown at the top of the installation page.
Start an aws-vault exec
session for the profile corresponding to the environment you want to connect to,
then navigate to the environment's directory and run terraform init
to create a local terraform state linking to the shared environment state.
You are now ready to start running terraform commands on the chosen environment.
The database is an RDS instance running on isolated subnets. This means there is no way to directly connect to it. Instead, we use one bastion EC2 instance per availability zone acting as an SSM agent. The instructions here assume that all infrastructure is deployed in the region eu-west-2 (London). If the infrastructure is in a different region, use that instead in each link and command.
Once you have followed the set up above, to connect to the database you will need:
- The id of bastion you intend to use - london ec2 instances
- The endpoint of the database - london databases
- The password for the database is stored in secrets - london secrets
First, start an aws-vault exec
session for the profile corresponding to the environment you want to connect to,
then run this command:
aws ssm start-session --region eu-west-2 --target <bastion id> --document-name AWS-StartPortForwardingSessionToRemoteHost --parameters host="<database endpoint>",portNumber="5432",localPortNumber="5432"
You should see something similar to:
Starting session with SessionId: 12a3456bcdefghi789
Port 5432 opened for sessionId 12a3456bcdefghi789.
Waiting for connections...
Leave this terminal open and running, and then you should be able to connect to the database from your machine using this connection string:
postgresql://postgres:<password>@localhost:5432/prsdb
If you are using DataGrip or Intellij's database tab and get a "Connection failed" error, instead of typing in the connection string try allowing it to write its own connection string using the settings:
- Host: localhost
- Port: 5432
- Authentication: User and Password
- User: postgres
- Password:
- Database: prsdb
When you start a connection, you will see a confirmation in the terminal window:
Connection accepted for session [12a3456bcdefghi789]
Alternatively, from the project root you can run:
./scripts/ssm_db_connect.ps1 <environment name>
in powershell, or./scripts/ssm_db_connect.sh <environment name>
in bash
This will start the port forwarding session, and copy the database password to your clipboard. You can then connect to the database as set out above.
After modifying the terraform files, you can run terraform fmt --recursive
from the root of the repository to format all the files.
Run terraform plan
to see what changes will be made, and if everything looks correct put up a PR for the changes.
The terraform will applied when your PR is merged.
If there are problems, it may be necessary to make updates manually from the terminal using terrform apply
, but we should aim to apply terraform via the git pipeline where possible.
We have javascript testing setup, initially this is for testing cloudfront functions.
Tests can be run using the "npm test" command from the root of the repository.
In order to interact with the environment in an AWS account you will need make sure you have a developer profile set up
for that account in your ~/.aws/config
. You should also update profile-template
with a template for the developer role.
You can use aws-vault
to create a terminal session with the appropriate credentials.
- Create the new folder
terraform/<environment name>
and copy the contents ofterraform/environment_template
into it - In
terraform/<environment name>/backend
remove the.template
from the end of the file name and replace all instances of the string<environment name>
with your actual environment name. - Leave the block that starts with
backend "s3"
near the top of the file commented out cd
intoterraform/<environment name>/backend
and runterraform init
followed byterraform plan
. Among other things the plan should show the creation of an s3 bucket calledprsdb-tfstate-<environment name>
and a dynamoDB table namedtfstate-lock-<environment name>
. If everything looks correct with the plan output, runterraform apply
.- After the
apply
step completes successfully and you can see the s3 bucket and dynamoDB table in the aws console, uncomment thebackend "s3"
block and runterraform init
- You should be prompted to move the terraform state to the remote backend. Once this is done the terraform state is successfully bootstrapped for the new environment
- In your new
terraform/<environment name>
folder, you can now also remove the.template
from the end of all of the filenames, replace all instances of the string<environment name>
with your actual environment name, and more generally look throughmain.tf
for any sets of<>
that require environment-specific input, e.g. the list of allowed IP addresses.
- Before we can create the environment as a whole, we must first create the initial networking infrastructure, and then request the DNS names and certificates from MHCLG.
- One of the files in your new
terraform/<environment name>
folder will beterraform.tfvars
. Check that this contains the linessl_certs_created = false
- Once this is done,
cd
intoterraform/<environment name>
and runterraform init
followed byterraform plan -target module.networking -target module.frontdoor
. If the output of the plan looks correct, runterraform apply -target module.networking -target module.frontdoor
to bring up the networking for the new environment. - Use the values in the terraform output to complete the request to MHCLG for changes to their DNS records
-
Use the terraform output to complete a copy of the 'DNS Change Request Form -v2.xlsx' file in the root of the repository as follows:
- For each item in the
cloudfront_certificate_validation
andload_balancer_certificate_validation
blocks of the output:- add a row to the table where:
- 'Change' Type is 'Add'
- 'Requested by' is your name
- 'Record Type' is the value from
resource_record_type
- 'Domain' is either
test.communities.gov.uk
orservice.gov.uk
, whichever appears as part of the value indomain_name
- 'Name' is the value from
resource_record_name
- 'Content' is the value from
resource_record_value
- 'TTL' is '1 hr'
- 'Proxy status' is 'DNS only'
- 'Additional comment or Reason for this change' is 'Setting up environment for the PRS Database'
- add a row to the table where:
- For each item in the
cloudfront_certificate_validation
block of the output add an additional row to the table where:- 'Change' Type is 'Add'
- 'Requested by' is your name
- 'Record Type' is 'CNAME'
- 'Domain' is either
test.communities.gov.uk
orservice.gov.uk
, whichever appears as part of the value indomain_name
- 'Name' is the value from
domain name
- 'Content' is the value from
cloudfront_dns_name
- 'TTL' is '1 hr'
- 'Proxy status' is 'DNS only'
- 'Additional comment or Reason for this change' is 'Setting up environment for the PRS Database'
- For each item in the
load_balancer_certificate_validation
block of the output add an additional row to the table where:- 'Change' Type is 'Add'
- 'Requested by' is your name
- 'Record Type' is 'CNAME'
- 'Domain' is either
test.communities.gov.uk
orservice.gov.uk
, whichever appears as part of the value indomain_name
- 'Name' is the value from
domain name
- 'Content' is the value from
load_balancer_dns_name
- 'TTL' is '1 hr'
- 'Proxy status' is 'DNS only'
- 'Additional comment or Reason for this change' is 'Setting up environment for the PRS Database'
- For each item in the
-
Once the spreadsheet is completed, create a service now request with MHCLG using the general 'Request' option with the following details:
- 'What is it that you require?' --> "Creation of DNS records"
- 'Why do you require it?' --> "We are setting up the environment for the new PRS Database in AWS. As part of this we need DNS records for the sub-domains and associated certificates. These subdomains were approved by TDA on 21/08/24. The service owner for the project is ."
- And add the completed spreadsheet as an attachment to the request
In order to use the prsdb web service once it's been deployed to these domains, One-Login needs to be updated to know about the new environment. Get the One-Login admin to add these for each of the cloudfront domains used above:
- Redirect URIs:
- /login/oauth2/code/one-login
- /login/oauth2/code/one-login-id
- Post log out redirect URIs:
- /signout
- We also need to create a number of pre-requisite resources, and then initial task definition before we can bring up the full ECS service.
- Make sure that the
terraform.tfvars
file that was created in the previous step contains the linetask_definition_created = false
- Still inside your
terraform/<environment name>
folder, runterraform plan
. - If the output of the plan looks correct, run
terraform apply
to create the pre-requisite resources. - Terraform will create the webapp secrets and SSM parameters but (apart from those related to the database and Redis) will not populate them. Ask the team lead where you can find the appropriate values, and then populate them via the AWS console.
- For Notify, this will involve creating a new API Key to use for the new environment.
- You will also need to add a new environment secret in Github called
ALARMS_EMAIL
with the email address to use for alarms in the new environment.
- To create the task definition, in
terraform/<environment name>/ecs_task_definition
, if you haven't already, remove the.template
from the end of any file names in the folder, and replace all instances of the string<environment name>
with your actual environment name. - In
terraform/<environment name>/ecs_task_definition/terraform.tfvars
, set thefile_upload_created
variable tofalse
. - Next, cd into
terraform/<environment name>/ecs_task_definition
and runterraform init
followed byterraform plan
. This should show you one resource being created - the task definition. If the output looks correct, runterraform apply
.
- In
terraform.tfvars
setssl_certs_created
totrue
andtask_definition_created
totrue
. - Then run
terraform plan
. If the output looks correct, runterraform apply
to bring up the environment. - Finally, in
terrafrom/<environment name>/ecs_task_definition
, set thefile_upload_created
variable totrue
, and runterraform plan
followed byterraform apply
to update the environment variables in the task definition.
This repository uses tfsec to scan the Terraform code for potential security issues. It can be run using Docker
docker run --pull=always --rm -it -v "$(pwd):/src" aquasec/tfsec /src
Individual rules can be ignored with a comment on the line above with the form tfsec:ignore:<rule-name>
e.g. tfsec:ignore:aws-dynamodb-enable-at-rest-encryption
.