This repo is part of a multi-part guide that shows how to configure and deploy the example.com reference architecture described in Google Cloud security foundations guide (PDF). The following table lists the parts of the guide.
0-bootstrap | Bootstraps a Google Cloud organization, creating all the required resources and permissions to start using the Cloud Foundation Toolkit (CFT). This step also configures a CI/CD pipeline for foundations code in subsequent stages. |
1-org | Sets up top level shared folders, monitoring and networking projects, and organization-level logging, and sets baseline security settings through organizational policy. |
2-environments | Sets up development, non-production, and production environments within the Google Cloud organization that you've created. |
3-networks | Sets up base and restricted shared VPCs with default DNS, NAT (optional), Private Service networking, VPC service controls, on-premises Dedicated Interconnect, and baseline firewall rules for each environment. It also sets up the global DNS hub. |
4-projects (this file) | Sets up a folder structure, projects, and application infrastructure pipeline for applications, which are connected as service projects to the shared VPC created in the previous stage. |
5-app-infra | Deploy a simple Compute Engine instance in one of the business unit projects using the infra pipeline set up in 4-projects. |
For an overview of the architecture and the parts, see the terraform-example-foundation README.
The purpose of this step is to set up the folder structure, projects, and infrastructure pipelines for applications that are connected as service projects to the shared VPC created in the previous stage.
For each business unit, a shared infra-pipeline
project is created along with Cloud Build triggers, CSRs for application infrastructure code and Google Cloud Storage buckets for state storage.
This step follows the same conventions as the foundation pipeline deployed in 0-bootstrap.
The Cloud Build SA used by this pipeline can impersonate the project SA by enabling the enable_cloudbuild_deploy
flag and necessary roles can be granted to this SA via sa_roles
as shown in this example.
This pipeline can be utilized for deploying resources in projects across development/non-production/production with granular permissions.
-
0-bootstrap executed successfully.
-
1-org executed successfully.
-
2-environments executed successfully.
-
3-networks executed successfully.
-
Obtain the value for the
access_context_manager_policy_id
variable.gcloud access-context-manager policies list --organization YOUR_ORGANIZATION_ID --format="value(name)"
-
For the manual step described in this document, you need Terraform version 0.13.7 to be installed.
Note: Make sure that you use the same version of Terraform throughout this series. Otherwise, you might experience Terraform state snapshot lock errors.
-
Obtain the values for the
perimeter_name
for each environment variable.gcloud access-context-manager perimeters list --policy ACCESS_CONTEXT_MANAGER_POLICY_ID --format="value(name)"
Note: If you have more than one service perimeter for each environment, you can also get the values from the
restricted_service_perimeter_name
output from each of the3-networks
environments.If you are using Cloud Build you can also search for the values in the outputs from the build logs:
gcloud builds list \ --project=YOUR_CLOUD_BUILD_PROJECT_ID \ --filter="status=SUCCESS \ AND source.repoSource.repoName=gcp-networks \ AND substitutions.BRANCH_NAME=development" \ --format="value(id)"
Use the result of this command as the
BUILD_ID
value in the next command:gcloud builds log BUILD_ID \ --project=YOUR_CLOUD_BUILD_PROJECT_ID | \ grep "restricted_service_perimeter_name = "
Change the
BRANCH_NAME
fromdevelopment
tonon-production
orproduction
for the other two service perimeters.
Please refer to troubleshooting if you run into issues during this step.
Note: You need to set variable enable_hub_and_spoke
to true
to be able to use the Hub-and-Spoke architecture detailed in the Networking section of the google cloud security foundations guide.
Note: If you are using MacOS, replace cp -RT
with cp -R
in the relevant
commands. The -T
flag is needed for Linux, but causes problems for MacOS.
- Clone repo.
gcloud source repos clone gcp-projects --project=YOUR_CLOUD_BUILD_PROJECT_ID
- Change freshly cloned repo and change to non-main branch.
git checkout -b plan
- Copy contents of foundation to new repo.
cp -RT ../terraform-example-foundation/4-projects/ .
- Copy Cloud Build configuration files for Terraform.
cp ../terraform-example-foundation/build/cloudbuild-tf-* .
- Copy Terraform wrapper script to the root of your new repository.
cp ../terraform-example-foundation/build/tf-wrapper.sh .
- Ensure wrapper script can be executed.
chmod 755 ./tf-wrapper.sh
- Rename
common.auto.example.tfvars
tocommon.auto.tfvars
and update the file with values from your environment and bootstrap. See any of the business unit envs folders README.md files for additional information on the values in thecommon.auto.tfvars
file. - Rename
shared.auto.example.tfvars
toshared.auto.tfvars
and update the file with values from your environment and bootstrap. See any of the business unit shared envs folders README.md files for additional information on the values in theshared.auto.example.tfvars
. - Rename
development.auto.example.tfvars
todevelopment.auto.tfvars
and update the file with theperimeter_name
that starts withsp_d_shared_restricted
. - Rename
non-production.auto.example.tfvars
tonon-production.auto.tfvars
and update the file with theperimeter_name
that starts withsp_n_shared_restricted
. - Rename
production.auto.example.tfvars
toproduction.auto.tfvars
and update the file with theperimeter_name
that starts withsp_p_shared_restricted
. - Rename
access_context.auto.example.tfvars
toaccess_context.auto.tfvars
and update the file with theaccess_context_manager_policy_id
. - You need to manually plan and apply only once the
business_unit_1/shared
environment sincedevelopment
,non-production
, andproduction
depend on it.- Run
cd ./business_unit_1/shared/
. - Update
backend.tf
with your bucket name from the 0-bootstrap step. - Run
terraform init
. - Run
terraform plan
and review output. - Run
terraform apply
. - Run
terraform output cloudbuild_sa
to get the cloud build service account from the apply step. - If you would like the bucket to be replaced by cloud build at run time, change the bucket name back to
UPDATE_ME
- Run
- Once you have done the instructions for the
business_unit_1
, you need to repeat same steps forbusiness_unit_2
folder. - Rename
business_unit_1.auto.example.tfvars
tobusiness_unit_1.auto.tfvars
and update the file with theapp_infra_pipeline_cloudbuild_sa
which is the output ofcloudbuild_sa
frombusiness_unit_1/shared
steps. - Rename
business_unit_2.auto.example.tfvars
tobusiness_unit_2.auto.tfvars
and update the file with theapp_infra_pipeline_cloudbuild_sa
which is the output ofcloudbuild_sa
frombusiness_unit_2/shared
steps. - Commit changes.
git add . git commit -m 'Your message'
- Push your plan branch to trigger a plan for all environments. Because the
plan branch is not a named environment branch, pushing your plan
branch triggers terraform plan but not terraform apply.
git push --set-upstream origin plan
- Review the plan output in your Cloud Build project https://console.cloud.google.com/cloud-build/builds?project=YOUR_CLOUD_BUILD_PROJECT_ID
- Merge changes to production. Because this is a named environment branch,
pushing to this branch triggers both terraform plan and terraform apply.
git checkout -b production git push origin production
- Review the apply output in your Cloud Build project. https://console.cloud.google.com/cloud-build/builds?project=YOUR_CLOUD_BUILD_PROJECT_ID
- After production has been applied, apply development.
- Merge changes to development. Because this is a named environment branch,
pushing to this branch triggers both terraform plan and terraform apply.
git checkout -b development git push origin development
- Review the apply output in your Cloud Build project https://console.cloud.google.com/cloud-build/builds?project=YOUR_CLOUD_BUILD_PROJECT_ID
- After development has been applied, apply non-production.
- Merge changes to non-production. Because this is a named environment branch,
pushing to this branch triggers both terraform plan and terraform apply.
git checkout -b non-production git push origin non-production
- Review the apply output in your Cloud Build project. https://console.cloud.google.com/cloud-build/builds?project=YOUR_CLOUD_BUILD_PROJECT_ID
- You can now move to the instructions in the step 5-app-infra.
- Clone the repo you created manually in 0-bootstrap.
git clone <YOUR_NEW_REPO-4-projects>
- Navigate into the repo and change to a non-production branch. All subsequent
steps assume you are running them from the gcp-environments directory. If
you run them from another directory, adjust your copy paths accordingly.
cd YOUR_NEW_REPO_CLONE-4-projects git checkout -b plan
- Copy contents of foundation to new repo.
cp -RT ../terraform-example-foundation/4-projects/ .
- Copy the Jenkinsfile script to the root of your new repository.
cp ../terraform-example-foundation/build/Jenkinsfile .
- Update the variables located in the
environment {}
section of theJenkinsfile
with values from your environment:_TF_SA_EMAIL _STATE_BUCKET_NAME _PROJECT_ID (the cicd project id)
- Copy Terraform wrapper script to the root of your new repository.
cp ../terraform-example-foundation/build/tf-wrapper.sh .
- Ensure wrapper script can be executed.
chmod 755 ./tf-wrapper.sh
- Rename
common.auto.example.tfvars
tocommon.auto.tfvars
and update the file with values from your environment and bootstrap. - Rename
shared.auto.example.tfvars
toshared.auto.tfvars
and update the file with values from your environment and bootstrap. - Rename
development.auto.example.tfvars
todevelopment.auto.tfvars
and update the file with theperimeter_name
that starts withsp_d_shared_restricted
. - Rename
non-production.auto.example.tfvars
tonon-production.auto.tfvars
and update the file with theperimeter_name
that starts withsp_n_shared_restricted
. - Rename
production.auto.example.tfvars
toproduction.auto.tfvars
and update the file with theperimeter_name
that starts withsp_p_shared_restricted
. - Rename
access_context.auto.example.tfvars
toaccess_context.auto.tfvars
and update the file with theaccess_context_manager_policy_id
. - You need to manually plan and apply only once the
business_unit_1/shared
environment sincedevelopment
,non-production
, andproduction
depend on it.- Run
cd ./business_unit_1/shared/
. - Update
backend.tf
with your bucket name from the 0-bootstrap step. - Run
terraform init
. - Run
terraform plan
and review output. - Run
terraform apply
. - Run
terraform output cloudbuild_sa
to get the cloud build service account from the apply step. - If you would like the bucket to be replaced by cloud build at run time, change the bucket name back to
UPDATE_ME
- Run
- Once you have done the instructions for the
business_unit_1
, you need to repeat same steps forbusiness_unit_2
folder. - Rename
business_unit_1.auto.example.tfvars
tobusiness_unit_1.auto.tfvars
and update the file with theapp_infra_pipeline_cloudbuild_sa
which is the output ofcloudbuild_sa
frombusiness_unit_1/shared
steps. - Rename
business_unit_2.auto.example.tfvars
tobusiness_unit_2.auto.tfvars
and update the file with theapp_infra_pipeline_cloudbuild_sa
which is the output ofcloudbuild_sa
frombusiness_unit_2/shared
steps. - Commit changes.
git add . git commit -m 'Your message'
- Push your plan branch.
git push --set-upstream origin plan
- Assuming you configured an automatic trigger in your Jenkins Master (see Jenkins sub-module README), this will trigger a plan. You can also trigger a Jenkins job manually. Given the many options to do this in Jenkins, it is out of the scope of this document see Jenkins website for more details.
- Review the plan output in your Master's web UI.
- Merge changes to production branch.
git checkout -b production git push origin production
- Review the apply output in your Master's web UI (you might want to use the option to "Scan Multibranch Pipeline Now" in your Jenkins Master UI).
- After production has been applied, apply development.
- Merge changes to development branch.
git checkout -b development git push origin development
- Review the apply output in your Master's web UI (you might want to use the option to "Scan Multibranch Pipeline Now" in your Jenkins Master UI).
- After development has been applied, apply non-production.
- Merge changes to non-production branch.
git checkout -b non-production git push origin non-production
- Review the apply output in your Master's web UI (you might want to use the option to "Scan Multibranch Pipeline Now" in your Jenkins Master UI).
- Change into 4-projects folder.
- Run
cp ../build/tf-wrapper.sh .
- Run
chmod 755 ./tf-wrapper.sh
. - Rename
common.auto.example.tfvars
tocommon.auto.tfvars
and update the file with values from your environment and bootstrap. - Rename
shared.auto.example.tfvars
toshared.auto.tfvars
and update the file with values from your environment and bootstrap. - Rename
development.auto.example.tfvars
todevelopment.auto.tfvars
and update the file with theperimeter_name
that starts withsp_d_shared_restricted
. - Rename
non-production.auto.example.tfvars
tonon-production.auto.tfvars
and update the file with theperimeter_name
that starts withsp_n_shared_restricted
. - Rename
production.auto.example.tfvars
toproduction.auto.tfvars
and update the file with theperimeter_name
that starts withsp_p_shared_restricted
. - Rename
access_context.auto.example.tfvars
toaccess_context.auto.tfvars
and update the file with theaccess_context_manager_policy_id
. - Update
backend.tf
with your bucket from the bootstrap step.You can runfor i in `find -name 'backend.tf'`; do sed -i 's/UPDATE_ME/<YOUR-BUCKET-NAME>/' $i; done
terraform output gcs_bucket_tfstate
in the 0-bootstrap folder to obtain the bucket name.
We will now deploy each of our environments(development/production/non-production) using this script. When using Cloud Build or Jenkins as your CI/CD tool each environment corresponds to a branch is the repository for 4-projects step and only the corresponding environment is applied. Environment shared must be applied first because development, non-production, and production depend on it.
To use the validate
option of the tf-wrapper.sh
script, please follow the instructions in the Install Terraform Validator section and install version v0.4.0
in your system. You will also need to rename the binary from terraform-validator-<your-platform>
to terraform-validator
and the terraform-validator
binary must be in your PATH
.
- Run
./tf-wrapper.sh init shared
. - Run
./tf-wrapper.sh plan shared
and review output. - Run
./tf-wrapper.sh validate shared $(pwd)/../policy-library <YOUR_CLOUD_BUILD_PROJECT_ID>
and check for violations. - Run
./tf-wrapper.sh apply shared
. - Rename
business_unit_1.auto.example.tfvars
tobusiness_unit_1.auto.tfvars
and update the file with theapp_infra_pipeline_cloudbuild_sa
which is the output ofcloudbuild_sa
frombusiness_unit_1/shared
steps. - Rename
business_unit_2.auto.example.tfvars
tobusiness_unit_2.auto.tfvars
and update the file with theapp_infra_pipeline_cloudbuild_sa
which is the output ofcloudbuild_sa
frombusiness_unit_2/shared
steps. - Run
./tf-wrapper.sh init production
. - Run
./tf-wrapper.sh plan production
and review output. - Run
./tf-wrapper.sh validate production $(pwd)/../policy-library <YOUR_CLOUD_BUILD_PROJECT_ID>
and check for violations. - Run
./tf-wrapper.sh apply production
. - Run
./tf-wrapper.sh init non-production
. - Run
./tf-wrapper.sh plan non-production
and review output. - Run
./tf-wrapper.sh validate non-production $(pwd)/../policy-library <YOUR_CLOUD_BUILD_PROJECT_ID>
and check for violations. - Run
./tf-wrapper.sh apply non-production
. - Run
./tf-wrapper.sh init development
. - Run
./tf-wrapper.sh plan development
and review output. - Run
./tf-wrapper.sh validate development $(pwd)/../policy-library <YOUR_CLOUD_BUILD_PROJECT_ID>
and check for violations. - Run
./tf-wrapper.sh apply development
.
If you received any errors or made any changes to the Terraform config or terraform.tfvars
you must re-run ./tf-wrapper.sh plan <env>
before running ./tf-wrapper.sh apply <env>
.