HowTo: Do better deployments

Edit on GitHub

Are you about to roll out an important feature to your staging or production environment and want to ensure that everything will work out right? Or you are encountering behavior in your application when it is deployed that does not seem right, and you are wondering how to best debug it? This document provides tips that can help you avoid surprises and help you prepare your project optimally for deplayment and build a local development setup with which you can debug more effectively.

Prerequisites

Read access to your codebase

Simulate your application

To simulate your application behavior and how it looks when deployed to the staging or production environment, bootstrap the deploy.yml files used by those environments. The following sections show what needs to be done.

Prepare your local hosts file

In these deploy.yml files, you specify the actual endpoint names that determine the URLs under which your environment is reachable. To work locally, point your DNS names to your local development environment by adding host entries in your local /etc/hosts file.

Verification

Make sure that all endpoints in your deploy.*.yml file are referenced there and point to 127.0.0.1.

Bootstrap with deploy.yml

For development purposes, your project has different deploy.yml files. However, to simulate the app’s behavior, you can use the deploy.yml files used during deployment to staging and production environments.

You must have the following deploy.yml files in your project (it may vary, depending on the total quantity of your environments):

  • deploy.(PROJECT_NAME)-prod.yml
  • deploy.(PROJECT_NAME)-staging.yml

PROJECT_NAME represents the name of your project.

Bootstrap your deploy.yml:

docker/sdk boot DEPLOY_FILE.yml && docker/sdk up

Replace DEPLOY_FILE.yml with the YML file of your choice.

It starts up your application, which is reachable through its staging and production URLs and behaves just like it would in your PaaS environments. This setup shows whether your application builds correctly with the deploy files used in the PaaS pipelines and lets you check out the look and feel of your application more authentically.

Ingest staging or production data

Mind the database load

Creating a database dump can cause a significant load on your database. If you are creating a database dump from a production environment, ensure you are doing so by using a read replica (if available). You can check whether you have a read replica for your production database by searching for RDS in the AWS console. This lets you find your read replica listed next to your production RDS database. You can obtain the host address there.

By ingesting the data in your staging or production database, you can go even one step further and bring your local environment even closer to its staging or production form. You can easily create a dump of your staging or production database by connecting to the RDS instance while having your VPN connected to the respective environment. For this, you need two things:

  • Your RDS instance URL.
  • Your DB credentials.

You can obtain all these things by logging in to the AWS console and searching for “Parameter Store”. Ensure that you have selected the right AWS region before you search. To list all the Parameter Store entries for the DB, you can enter “DB” in the search. You are looking for the following parameters:

  • /codebuild/base_task_definition/SPRYKER_DB_HOST
  • /codebuild/base_task_definition/SPRYKER_DB_ROOT_USERNAME
  • /codebuild/base_task_definition/SPRYKER_DB_ROOT_PASSWORD

With this information, you can connect to the database from any SQL client and create a database dump which you can then import locally. After you have imported the data, don’t forget to publish events so that all the data gets imported to Redis and Elastic Search as well. You can use the following command to achieve that.

command publish:trigger-events