logo
Menu

Streamline AWS Deployments with GitLab CI and Terraform

A step-by-step guide for deploying infrastructure using CI/CD.

Kaizad Wadia
Amazon Employee
Published Dec 18, 2023
Last Modified Mar 18, 2024
Deploying infrastructure to AWS manually can be time consuming and prone to errors. Mistakes can easily happen because manual processes rely on humans repeating the same tasks without validation checks, leading to configuration errors, inconsistencies, and drift from the desired state. In this post, we'll go over how to set up a CI/CD pipeline for automating AWS deployments using GitLab and Terraform.

Overview

First, we'll write a Terraform script to deploy AWS resources and a matching .gitlab-ci.yml file to execute the script on commit.
We'll configure the necessary permissions in AWS IAM to allow GitLab to create infrastructure. We'll also create an S3 bucket and DynamoDB table to store remote Terraform state.
Finally, we'll test the pipeline by deploying sample infrastructure.
By following this guide, you'll learn infrastructure-as-code best practices for GitLab and Terraform and have an automated deployment pipeline to AWS ready for your infrastructure repos. Let's get started on setting up smooth AWS deployments with GitLab and Terraform.

Prerequisites

You need an AWS account and a GitLab (free trial) account. Note that the steps in this tutorial will be different if you choose to use a self-managed GitLab account. This difference is mainly applicable to the second step.
About
✅ AWS LevelIntermediate - 200
⏱ Time to complete45 minutes
💰 Cost to completeFree Tier Eligible
🧩 PrerequisitesAWS Account and GitLab Account
📢 FeedbackAny feedback, issues, or just a 👍 / 👎
⏰ Last Updated2023-12-18

Walkthrough

Build the Repository

We will start by creating the code repository on GitLab that is to be deployed to AWS. In GitLab, create a new repository to store the Terraform script by clicking on "New project" on the top right. Then click on "Create blank project". For the project name, enter "My Terraform Queue" and for the group (the first part of the project URL), we can select our username (for me that would be kaizadwadia). The project slug should be generated automatically while for the rest of the fields, the default options are sufficient.
Create GitLab Repo
Once we click on "Create Project" we should be navigated to the home page of our repository. To make changes to the repository we will use the Web IDE, which is essentially a web-based VSCode. To open that, we can click on the "Edit" dropdown on the right side of the screen to the left of the "Code" button, and select "Web IDE". The IDE should open in another browser tab.
We will now create our Terraform file. Click on the "New File" icon which is on the top left-hand side of the screen and should appear when the cursor is over the repository name. This can also be done by right clicking on the left side of the screen underneath the "README.md" file and clicking "New File".
Web IDE
Then we can give our file a name. I chose "main.tf" for the file name. Upon pressing enter, the file will be opened. Let's add the code that creates the SQS Queue by copying the following code to the "main.tf" file.
Now we can save this code by pressing Ctrl + s (Cmd + s on Mac) and commit the code to the main branch by going to the source control tab on the sidebar (third from the top) and clicking "Commit to 'main'". If a warning pops up, we can override it by clicking "Continue" instead of creating a new branch.
Commit to main
We will come back to modify this code later, so let's leave this tab open. For now, let's move on to allowing GitLab to modify resources in our AWS account.

Configure an Identity Provider in IAM

We will now configure GitLab as an identity provider for IAM. This allows IAM to recognize GitLab as a trusted entity that we can assign permissions via an IAM Role.
To do this, navigate to IAM in the AWS Management Console, and then click on "Identity Providers" on the sidebar under "Access Management". After that, click on the button saying "Add provider" on the top right-hand side. Once there, enter the information as displayed in the below image.
Adding IdP
Make sure that the option for OpenID Connect is selected on the top, this is because OpenID Connect allows you to configure trust relationships with third party accounts including but not limited to GitLab. Also ensure that the provider URL is GitLab's root URL, https://gitlab.com as well as the Audience. The provider URL is the URL from which the OpenID configuration is obtained. You can view the OpenID configuration for GitLab yourself by navigating to gitlab.com/.well-known/openid-configuration. Note that if you are using a self-managed instance of GitLab (that is publicly available on the internet) you would need to use the root URL of that instance instead of gitlab.com. In this case, check that your instance supports OpenID connect by navigating to the OpenID configuration URL using your instance's root domain.
Once you have completely filled out the information, you can click on "Get thumbprint" to retrieve the thumbprint from the URL you supplied. The OIDC thumbprint page in the AWS documentation describes how this thumbprint is obtained. Click on "Add provider". We have now configured GitLab as an Identity Provider in our AWS Account.

Create an IAM Role for GitLab CI

To create an IAM Role that our CI/CD pipeline can assume to make deployments, click on "Roles" under "Access Management" in the IAM page and then click on the "Create role" button on the top right of the page. In the first step, configuring the trusted entity means providing access to our identity provider we created before. For this, select on "Web identity", which should allow you to select "gitlab.com" as the identity provider from the dropdown menu and "https://gitlab.com" as the audience, similar to the below image.
Trusted Entities Creation
Once that is done, clicking "next" should allow us to select IAM policies which determine what permissions GitLab should have. For the purposes of this tutorial, we will provide full access only to three services: Amazon S3, DynamoDB, and SQS by attaching the following policies:
  • AmazonS3FullAccess
  • AmazonDynamoDBFullAccess
  • AmazonSQSFullAccess
Once we have selected the appropriate policies, we can move on to the next step, which allows us to give the role a name and description. The role's name can be "GitLabRole" and a fair description of the role could be "An IAM Role to give GitLab CI access to deploy to AWS". We can now scroll down to the bottom and click on "Create role".
Once the role is created, navigate to it in the console under "Roles" by searching for the role name before clicking on the role itself. Once you are on the role's page, select the "Trust relationships" tab and click on "Edit trust policy" on the right side of the page. Edit the condition in the trust policy to something like this:
Replace {GITLAB_GROUP} with your username (eg. kaizadwadia), {GITLAB_PROJECT} with your project slug (ie. my-terraform-queue), and {GITLAB_BRANCH} with main. This will ensure that only pipelines running on the main branch of the specific repository has permissions to assume this role. Moreover, note that while the "StringEquals" condition would work here, we can also use wildcards like "*" within the condition statement using the "StringLike" condition. This allows for more complex permissions assignments such as allowing any repository under your username to assume this role. The trust policy should now look something like this:
Upon clicking "Update Policy" the role should be ready for use by GitLab CI. Note the Role's ARN, as it will be used in a later step. To learn more about creating the trust policy, check out the GitLab documentation page.

Create the AWS Resources to Manage Terraform State

When creating resources with Terraform, files are created to manage the state of the infrastructure created by Terraform. Terraform state is a file that tracks the infrastructure that Terraform has created. It maps real-world resources to what is defined in your Terraform configuration files. The state file allows Terraform to determine what infrastructure needs to be created, updated, or destroyed based on the desired configuration.
State lock is a mechanism that prevents multiple Terraform commands from accessing and trying to modify the state at the same time. When you run a Terraform command, it will acquire a lock on the state so that no other commands can access it. This prevents race conditions where two commands try to update the state simultaneously in an unsafe way. Once the command finishes, it releases the lock so other commands can run.
We will create an S3 bucket to house the state file, and a DynamoDB Table to host the state locking mechanism. All resources will be created in the "eu-west-1" region for this tutorial. To create the S3 bucket, head over to the S3 page in the AWS Console and click "Create bucket" on the right side. Give the bucket a globally unique name, such as "terraform-state-bucket-XXXX", where XXXX is a random sequence of digits.
S3 Bucket
Now to create a DynamoDB Table, we can visit the DynamoDB page on the AWS Console, and click on "Create table" on the right side of the page. Make sure it is done in the same region as the S3 bucket. For the table name, we can give it "terraform-state-lock-table". We can call the Partition Key "LockID", and have it be a string.
DynamoDB Table
We can leave the rest as the default settings and click on "Create table". We have now completed the setup for state management. Now we will write the instructions in GitLab on how to deploy a Terraform script to AWS.

Write the Instructions for the CI/CD Pipeline

Before we write the instructions for the CI/CD Pipeline we need to give the IAM Role ARN to the Pipeline in the form of a secret variable so that our pipeline knows what role to assume. To do this, head over to the home page of the GitLab repository that we created in the first step, and click on "CI/CD" under the "Settings" dropdown menu on the very bottom of the sidebar. Then click "Expand" on the "Variables" section and click on "Add variable". Have the Key be "ROLE_ARN" and the value be the ARN of the Role that was saved in the third step.
GitLab Variable
All the other fields can be left as is and once "Add variable" is clicked on the bottom right-hand side, this variable is saved. Repeat this process once again for the region, saving it with the key, "AWS_DEFAULT_REGION" and the value should be the region code. For this tutorial, that would be the same region as before, "eu-west-1".
Variables
To write the instructions used by the CI/CD Pipeline to run our Terraform script we will create a .gitlab-ci.yml file in the root of our repository. See the first section on how to create and edit files using the Web IDE in GitLab. Once we have created the .gitlab-ci.yml file, we will add the following stage to it:
This YAML definition creates a stage called "deploy" that only runs when commited to the 'main' branch, using the Terraform Light Docker image. The magic is in the ID token which is the web identity token used to assume the IAM Role we created earlier in the third step. This is now provided in the environment variable, GITLAB_OIDC_TOKEN and available in the container to use. In the before_script, we are performing the configuration used in the process that assumes the IAM role with the web identity token and saving it under the AWS CLI profile called "oidc". Profiles in the AWS CLI allow you to configure and store credentials and settings for different AWS accounts or IAM users, keeping them separate and switching between them easily. Save the YAML file, and in the next section, we will bring everything together and deploy the code to AWS.

Deploy the Script

Now in the Web IDE, go back to the "main.tf" file we created earlier, and add the following code above the definition of the SQS Queue:
Make sure to replace "XXXX" with the actual name of the S3 bucket we created earlier. Note that the profile "oidc" is the same as the profile created by the pipeline, so that the Terraform script knows the role to assume while provisioning the resources. The same applies with the profile for the provider. Note that these two don't necessarily need to be the same, so you could use one role for accessing the backend and another for deploying the resources. The "key" parameter in the backend denotes the name of the file that is created to manage the Terraform state in the S3 bucket. The "encrypt" parameter determines whether SSE-S3 is used to encrypt the file in the bucket.
Once we have this written, commit this to the repository using the same method as in the first step. When we do so, our pipeline should start automatically upon committing our code to the main branch. We can see it through GitLab if we go back to our repository's home page and click on "Jobs" under the "Build" dropdown menu on the sidebar.
Deploy
If everything goes as planned we should see that our queue has been created in the pipeline's logs, as well as in the AWS Console if we navigate to the SQS page.

Clean Up

To clean up your AWS account, remember to delete the following resources:
  • The SQS Queue created by Terraform
  • The S3 bucket we created to store the Terraform state files
  • The DynamoDB Table called "terraform-state-lock-table"
  • The IAM Role, "GitLabRole" and Identity Provider, "gitlab.com"

Conclusion

And that's it! We've set up an automated AWS deployment pipeline using GitLab and Terraform best practices. Now any time you push changes to your infrastructure code, GitLab CI will trigger Terraform to deploy those changes to AWS.
In this post we covered configuring IAM permissions and roles for GitLab, creating S3 and DynamoDB resources to store remote state, writing a Terraform script to deploy infrastructure, setting up a .gitlab-ci.yml pipeline to execute on commits, and testing our automated deployments.
As you expand your infrastructure, be sure to continue using security best practices like least privilege IAM roles. And leverage GitLab's power with features like the Terraform template library.
To learn more about deploying to AWS using CI/CD, check out the following resources:

Any opinions in this post are those of the individual author and may not reflect the opinions of AWS.

Comments