Skip to content
👷 The Terragrunt website redesign is Work In Progress! 👷
For a list of outstanding TODOs see this.
To give feedback, click here.

Stacks

Motivation

Let’s say your infrastructure is defined across multiple OpenTofu/Terraform modules:

  • Directoryroot
    • Directorybackend-app
      • main.tf
    • Directoryfrontend-app
      • main.tf
    • Directorymysql
      • main.tf
    • Directoryredis
      • main.tf
    • Directoryvpc
      • main.tf

There is one unit to deploy a frontend-app, another to deploy a backend-app, another for the MySQL database, and so on.

To deploy such an environment, you’d have to manually run tofu/terraform apply in each of module, wait for it to complete, and then run tofu apply/terraform apply in the next subfolder. Moreover, you have to make sure you manually run tofu/terraform apply in the right module each time. The order in which they are applied can be important, especially if one module depends on another.

How do you avoid this tedious, error-prone and time-consuming process?

Stacks to the rescue

Terragrunt provides special tooling for operating on sets of units at once. Sets of units in Terragrunt are called stacks.

Right now, there is no special syntax for defining a stack as a single file, but that will change soon with the introduction of the Stacks RFC.

Regardless of whether you are using the new syntax or not, the core functionality provided by Stacks are the same. The new terragrunt.stack.hcl file is merely a shorthand for defining a stack, just like you could by placing units directly in a folder.

The run --all command

To solve the problem above, first convert the OpenTofu/Terraform modules into units. This is done simply by adding an empty terragrunt.hcl file within each root module folder.

  • Directoryroot
    • Directorybackend-app
      • main.tf
      • terragrunt.hcl
    • Directoryfrontend-app
      • main.tf
      • terragrunt.hcl
    • Directorymysql
      • main.tf
      • terragrunt.hcl
    • Directoryredis
      • main.tf
      • terragrunt.hcl
    • Directoryvpc
      • main.tf
      • terragrunt.hcl

Because you’ve created a directory of units, you’ve also implicitly created a stack!

Now, you can go into the root directory and deploy all the units within it by using the run --all command with apply:

Terminal window
cd root
terragrunt run --all apply

When you run this command, Terragrunt will recursively discover all the units under the current working directory, and run terragrunt apply on each of those units concurrently*.

Similarly, to undeploy all the OpenTofu/Terraform units, you can use the run --all command with destroy:

Terminal window
cd root
terragrunt run --all destroy

To see the currently applied outputs of all of the subfolders, you can use the run --all command with output:

Terminal window
cd root
terragrunt run --all output

Finally, if you make some changes to your project, you could evaluate the impact by using run --all command with plan:

Note: It is important to realize that you could get errors running run --all plan if you have dependencies between your projects and some of those dependencies haven’t been applied yet.

Ex: If unit A depends on unit B and unit B hasn’t been applied yet, then run --all plan will show the plan for B, but exit with an error when trying to show the plan for A.

Terminal window
cd root
terragrunt run --all plan

* Note that the units might run concurrently, but some units can be blocked from running until their dependencies are run.

If your units have dependencies between them, for example, you can’t deploy the backend-app until MySQL and valkey are deployed. You’ll need to express those dependencies in your Terragrunt configuration as explained in the next section.

Cutting down the file count

A simple way to reduce the file count from adding terragrunt.hcl files above is to move to remote OpenTofu/Terraform modules.

This is a best practice, especially in larger projects, as it allows you to pin specific versions of your modules and use that specific version across multiple units.

The rest of this documentation will assume you are following that best practice.

Passing outputs between units

Consider the following file structure:

  • Directoryroot
    • Directorybackend-app
      • terragrunt.hcl
    • Directorymysql
      • terragrunt.hcl
    • Directoryredis
      • terragrunt.hcl
    • Directoryvpc
      • terragrunt.hcl

Suppose that you wanted to pass in the VPC ID of the VPC that is created from the vpc unit in the folder structure above to the mysql unit as an input variable. Or that you wanted to pass in the subnet IDs of the private subnet that is allocated as part of the vpc unit.

You can use the dependency block to extract those outputs and use them as inputs to another unit.

For example, suppose the vpc unit outputs the ID under the name vpc_id. To access that output, you would specify in mysql/terragrunt.hcl:

mysql/terragrunt.hcl
dependency "vpc" {
config_path = "../vpc"
}
inputs = {
vpc_id = dependency.vpc.outputs.vpc_id
}

When you apply this unit, the output will be read from the vpc unit and passed in as an input to the mysql unit right before calling tofu apply/terraform apply.

You can also specify multiple dependency blocks to access the outputs of multiple units.

For example, in the above folder structure, you might want to reference the domain output of the redis and mysql units for use as inputs in the backend-app unit. To access those outputs, you would specify the following in backend-app/terragrunt.hcl:

backend-app/terragrunt.hcl
dependency "mysql" {
config_path = "../mysql"
}
dependency "redis" {
config_path = "../redis"
}
inputs = {
mysql_url = dependency.mysql.outputs.domain
redis_url = dependency.redis.outputs.domain
}

Note that each dependency block results in a relevant status in the Terragrunt DAG. This means that when you run run --all apply on a config that has dependency blocks, Terragrunt will not attempt to deploy the config until all the units referenced in dependency blocks have been applied. So for the above example, the order for the run --all apply command would be:

  1. Deploy the VPC

  2. Deploy MySQL and Redis in parallel

  3. Deploy the backend-app

If any of the units failed to deploy, then Terragrunt will not attempt to deploy the units that depend on them.

Note: Not all blocks can access outputs passed by dependency blocks. See the section on Configuration parsing order for more information.

Unapplied dependency and mock outputs

Terragrunt will return an error if the unit referenced in a dependency block has not been applied yet. This is because you cannot actually fetch outputs out of an unapplied unit, even if there are no resources being created in the unit.

This is most problematic when running commands that do not modify state (e.g run --all plan and run --all validate) on a completely new setup where no infrastructure has been deployed. You won’t be able to plan or validate a unit if you can’t determine the inputs. If the unit depends on the outputs of another unit that hasn’t been applied yet, you won’t be able to compute the inputs unless the dependencies are all applied.

Of course, in real life usage, you typically need the ability to run run --all validate or run --all plan on a completely new set of infrastructure.

To address this, you can provide mock outputs to use when a unit hasn’t been applied yet. This is configured using the mock_outputs attribute on the dependency block and it corresponds to a map that will be injected in place of the actual dependency outputs if the target config hasn’t been applied yet.

Using a mock output is typically the best solution here, as you typically don’t actually care that an accurate value is used for a given value at this stage, just that it will plan successfully. When you actually apply the unit, that’s when you want to be sure that a real value is used.

For example, in the previous scenario with a mysql unit and vpc unit, suppose you wanted to mock a value for the vpc_id during a run --all validate for the mysql unit.

You can specify that in mysql/terragrunt.hcl:

mysql/terragrunt.hcl
dependency "vpc" {
config_path = "../vpc"
mock_outputs = {
vpc_id = "mock-vpc-id"
}
}
inputs = {
vpc_id = dependency.vpc.outputs.vpc_id
}

You can now run validate on this config before the vpc unit is applied because Terragrunt will use the map {vpc_id = "mock-vpc-id"} as the outputs attribute on the dependency instead of erroring out.

What if you wanted to restrict this behavior to only the validate command? For example, you might not want to use the defaults for a plan operation because the plan doesn’t give you any indication of what is actually going to be created.

You can use the mock_outputs_allowed_terraform_commands attribute to indicate that the mock_outputs should only be used when running those OpenTofu/Terraform commands. So to restrict the mock_outputs to only when validate is being run, you can modify the above terragrunt.hcl file to:

mysql/terragrunt.hcl
dependency "vpc" {
config_path = "../vpc"
mock_outputs = {
vpc_id = "temporary-dummy-id"
}
mock_outputs_allowed_terraform_commands = ["validate"]
}
inputs = {
vpc_id = dependency.vpc.outputs.vpc_id
}

Note that indicating validate means that the mock_outputs will be used either with validate or with run --all validate.

You can also use skip_outputs on the dependency block to specify the dependency without pulling in the outputs:

mysql/terragrunt.hcl
dependency "vpc" {
config_path = "../vpc"
skip_outputs = true
}

When skip_outputs is used with mock_outputs, mocked outputs will be returned without attempting to load outputs from OpenTofu/Terraform.

This can be useful when you disable backend initialization (remote_state.disable_init) in CI for example.

mysql/terragrunt.hcl
dependency "vpc" {
config_path = "../vpc"
mock_outputs = {
vpc_id = "temporary-dummy-id"
}
skip_outputs = true
}

You can also use mock_outputs_merge_strategy_with_state on the dependency block to merge mocked outputs and real outputs:

mysql/terragrunt.hcl
dependency "vpc" {
config_path = "../vpc"
mock_outputs = {
vpc_id = "temporary-dummy-id"
new_output = "temporary-dummy-value"
}
mock_outputs_merge_strategy_with_state = "shallow"
}

If real outputs only contain vpc_id, dependency.outputs will contain a real value for vpc_id and a mocked value for new_output.

Dependencies between units

You can also specify dependencies explicitly. Consider the following file structure:

  • Directoryroot
    • Directorybackend-app
      • terragrunt.hcl
    • Directoryfrontend-app
      • terragrunt.hcl
    • Directorymysql
      • terragrunt.hcl
    • Directoryredis
      • terragrunt.hcl
    • Directoryvpc
      • terragrunt.hcl

Let’s assume you have the following dependencies between OpenTofu/Terraform units:

  • backend-app depends on mysql, redis, and vpc

  • frontend-app depends on backend-app and vpc

  • mysql depends on vpc

  • redis depends on vpc

  • vpc has no dependencies

You can express these dependencies in your terragrunt.hcl config files using a dependencies block. For example, in backend-app/terragrunt.hcl you would specify:

backend-app/terragrunt.hcl
dependencies {
paths = ["../vpc", "../mysql", "../redis"]
}

Similarly, in frontend-app/terragrunt.hcl, you would specify:

frontend-app/terragrunt.hcl
dependencies {
paths = ["../vpc", "../backend-app"]
}

Once you’ve specified these dependencies in each terragrunt.hcl file, Terragrunt will be able to perform updates respecting the DAG of dependencies.

For the example at the start of this section, the order of runs for the run --all apply command would be:

  1. Deploy the VPC

  2. Deploy MySQL and Redis in parallel

  3. Deploy the backend-app

  4. Deploy the frontend-app

Any error encountered in an individual unit during a run --all command will prevent Terragrunt from proceeding with the deployment of any dependent units.

To check all of your dependencies and validate the code in them, you can use the run --all validate command.

Visualizing the DAG

To visualize the dependency graph you can use the graph-dependencies command (similar to the terraform graph command).

The graph is output in DOT format. The typical program used to render this file format is GraphViz, but many web services are available that can do this as well.

Terminal window
terragrunt graph-dependencies | dot -Tsvg > graph.svg

The example above generates the following graph:

terragrunt graph-dependencies

Note that this graph shows the dependency relationship in the direction of the arrow, with the tip pointing to the dependency (e.g. frontend-app depends on backend-app).

For most commands, Terragrunt will run in the opposite direction, however (e.g. backend-app would be applied before frontend-app).

The exception to this rule is during the destroy (and plan -destroy) command, where Terragrunt will run in the direction of the arrow (e.g. frontend-app would be destroyed before backend-app).

Testing multiple units locally

If you are using Terragrunt to download remote OpenTofu/Terraform modules and all of your units have the source parameter set to a Git URL, but you want to test with a local checkout of the code, you can use the --source parameter to override that value:

Terminal window
cd root
terragrunt run --all plan --source /source/modules

If you set the --source parameter, the run --all command will assume that parameter is pointing to a folder on your local file system that has a local checkout of all of your OpenTofu/Terraform modules.

For each unit that is being processed via a run --all command, Terragrunt will:

  1. Read in the source parameter in that unit’s terragrunt.hcl file.
  2. Parse out the path (the portion after the double-slash).
  3. Append the path to the --source parameter to create the final local path for that unit.

For example, consider the following terragrunt.hcl file:

terragrunt.hcl
terraform {
source = "git::git@github.com:acme/infrastructure-modules.git//networking/vpc?ref=v0.0.1"
}

Running the following:

Terminal window
terragrunt run --all apply --source /source/infrastructure-modules

Will result in a unit with the configuration for the source above being resolved to /source/infrastructure-modules//networking/vpc.

Limiting run parallelism

By default, Terragrunt will not impose a limit on the number of units it executes when it traverses the dependency graph, meaning that if it finds 5 units without dependencies, it’ll run OpenTofu/Terraform 5 times in parallel, once in each unit.

Sometimes, this can create a problem if there are a lot of units in the dependency graph, like hitting a rate limit on a cloud provider.

To limit the maximum number of unit executions at any given time use the --parallelism [number] flag

Terminal window
terragrunt run --all apply --parallelism 4

Saving OpenTofu/Terraform plan output

A powerful feature of OpenTofu/Terraform is the ability to save the result of a plan as a binary or JSON file using the -out flag.

Terragrunt provides special tooling in run --all execution to ensure that the saved plan for a run --all against a stack has a corresponding entry for each unit in the stack in a directory structure that mirrors the stack structure.

To save plan against a stack, use the --out-dir flag (or TG_OUT_DIR environment variable) as demonstrated below:

Terminal window
$ terragrunt run --all plan --out-dir /tmp/tfplan
  • Directoryapp1
    • tfplan.tfplan
  • Directoryapp2
    • tfplan.tfplan
  • Directoryapp3
    • tfplan.tfplan
  • Directoryproject-2
    • Directoryproject-2-app1
      • tfplan.tfplan
Terminal window
$ terragrunt run --all --out-dir /tmp/tfplan apply

For planning a destroy operation, use the following commands:

Terminal window
terragrunt run --all --out-dir /tmp/tfplan plan -destroy
terragrunt run --all --out-dir /tmp/tfplan apply

To save plan in json format use --json-out-dir flag (or TG_JSON_OUT_DIR environment variable):

Terminal window
terragrunt run --all --json-out-dir /tmp/json plan
  • Directoryapp1
    • tfplan.json
  • Directoryapp2
    • tfplan.json
  • Directoryapp3
    • tfplan.json
  • Directoryproject-2
    • Directoryproject-2-app1
      • tfplan.json
Terminal window
terragrunt run --all --out-dir /tmp/all --json-out-dir /tmp/all plan
  • Directoryapp1
    • tfplan.json
    • tfplan.tfplan
  • Directoryapp2
    • tfplan.json
    • tfplan.tfplan
  • Directoryapp3
    • tfplan.json
    • tfplan.tfplan
  • Directoryproject-2
    • Directoryproject-2-app1
      • tfplan.json
      • tfplan.tfplan

To recap:

  • The plan for each unit in a stack is saved in the same hierarchy as the unit structure.
  • The file name for plan binaries are tfplan.tfplan and tfplan.json for plan JSON.
  • JSON plan files can’t be used with terragrunt run --all apply command, only binary plan files can be used.
  • Output directories can be combined which will lead to saving both binary and JSON plans.

Nested Stacks

Note that you can also have nested stacks.

For example, consider the following file structure:

  • Directoryroot
    • Directoryus-east-1
      • Directoryapp
        • terragrunt.hcl
      • Directorydb
        • terragrunt.hcl
    • Directoryus-west-2
      • Directoryapp
        • terragrunt.hcl
      • Directorydb
        • terragrunt.hcl

In this example, there’s the root stack, that contains all the infrastructure you’ve defined so far, and there’s also the us-east-1 and us-west-2 stacks, that contain the infrastructure for the app and db units in those regions.

You can run run --all commands at any depth of the stack to run the units in that stack and all of its children.

For example, to run all the units in the us-east-1 stack, you can run:

Terminal window
cd root/us-east-1
terragrunt run --all apply

Terragrunt will only include the units in the us-east-1 stack and its children in the queue of units to run (unless external dependencies are pulled in, as discussed in the run —all command section).

Generally speaking, this is the primary tool Terragrunt users use to control the blast radius of their changes. For the most part, it is the current working directory that determines the blast radius of a run --all command.

In addition to using your working directory to control what’s included in a run queue, you can also use flags like —include-dir and —exclude-dir to explicitly control what’s included in a run queue within a stack, or outside of it.

There are more flags that control the behavior of the run command, which you can find in the run docs.