What’s a Docker?

Docker is an open source technology used in app deployment. Whenever you develop an app, you do it with chosen operating system in mind (like a particular release of Linux), as well as a language of choice, associated tooling and libraries, which we can collectively refer to as dependencies. Whenever you run your app on some machine (for example on a rented cloud compute instance) you need to make sure that all of the dependencies are met for the app to run properly. Docker provides you an opportunity to do away with that headache by introducing a concept known as containers. Putting your app inside one is called containerization.

Core Docker Concepts

Container

Containers can be thought of as boxes where your app will run, isolated from the host’s operating system. Inside them you can freely define your dependencies and stop worrying about which host your app will run on. Thanks to the containers you can rest easy knowing that inside the box your app will have all the dependencies it needs to work properly. The host only needs to have Docker installed to launch your container with an app.

It is important to note that containers have their own isolated, minimal file systems inside, which house your app’s code. Due to this isolation containers also have their own set of ports, which often need to be configured to communicate with your local host.

Image

Whenever you spawn a container, it is spawned based on an image, which is another core Docker concept. It contains all the code and dependencies necessary to deploy your app inside a container.

Images are usually uploaded to image registries, which can be thought of as Github equivalents for images. They allow you to track the versioning of the image.

Some of the popular image registries are: Docker Hub, Google Container Registry (GCR), Amazon Elastic Container Registry (ECR) and Azure Container Registry (ACR). The last three are integrated with the most popular cloud providers and for a good reason - images are the current standard in app deployment in the cloud.

Dockerfile

Creating an image (commonly referred to as building) is performed by your local Docker installation based on a set of instructions in a so-called Dockerfile. These steps are defined by you and should not only copy the app code from your local machine to the image, but also install all the necessary tooling and libraries. They should also provide some basic information about the image, like which command to run upon container start up or which ports in the container should be exposed.

The filename “Dockerfile” is obligatory and is recognized by Docker as an instruction to build an image. Those instructions are defined in Docker specific syntax.

A Dockerfile is subject to the same version control system as other files in a project, which means that any changes in image setup can be easily tracked and documented.

Containerization Thought Map

In other words, if you want to put your app inside a container, the steps towards doing so are:

1) Define the dependencies for your app

2) Write a Dockerfile where those dependencies will be addressed

3) Build an image

4) Launch a container based on that image

Base Docker Commands

After you install Docker on your local machine, you should get familiar with some of the basic Docker commands to work with containers: 1) Build an image called image_name based on a Dockerfile in the current directory.

docker build -t image_name .

2) Launch a new container based on an image in an interactive mode (you will be able to interact with it from the terminal, which is very useful for debugging)

docker run -it --name container_name image_name

3) Launch an interactive shell to get inside the container and interact with the files and logs. Extremely useful in debugging.

docker exec -it container_name bash

4) Stop a running container. It will still be left on your system and you’ll be able to restart it.

docker stop container_name

5) Remove a container from your system. If it’s running it will be stopped first.

docker rm container_name

5) List the images currently present in your system

docker images

6) List currently running containers

docker ps

Let’s Build an App

Baduk.ninja

In this exemplary app containerization I will share a Dockerfile from baduk.ninja - a website dedicated to all aspiring Go players aiming to improve their skills. While the source code for this project is currently private, it can still be used as a valid example for building a simple image for an app.

Project Structure

The project can be summarised by the following directory hierarchy:

  • baduk.ninja
    • frontend
      • node_modules
      • src
      • public
      • package-lock.json
      • package.json
    • backend
      • src
      • Cargo.lock
      • Cargo.toml
    • sgfs
    • Dockerfile
    • run_services.sh
  • rust-goban-fork

    Defining Dependencies

    The frontend is written in JavaScript and uses React, with directory structure as expected by the npm package manager. It listens on port 80 and communicates with backend on port 4000.

The backend is written in Rust and is structured to work with cargo package manager. Aside from official library releases, it uses a fork of the goban library, which we will have to handle while writing a Dockerfile. It also uses some sgfs files from a separate directory.

In the main baduk.ninja directory we can also see a run_services.sh file, which will launch both fronted and backend. It will track the status of those two services and if at least one of them exits with an error, the entire process will, too. This will come in handy later, as run_services.sh will be the main entry script for our future containers. If either of our running services exits inside the container, we want the container itself to go down, too.

Image Concept

We will build a single image that will run both the frontend and backend of our app. Since we’re using ports 80 and 4000, our future containers will have to expose those two ports to communicate with the host.

In the image we are also looking for a way to build both the frontend and backend and use those executables in starting the services for an optimized performance.

Picking a Base Image

What is a Base Image?

It is important to note that when building an image we don’t do so from an absolute scratch. The first step is to pick a base image, which provides a basic file structure and some preinstalled tooling, and add our own changes on top of those. An example of a base image can be as simple as a minimal installation of Ubuntu, or a more robust Node or Rust base images. The last two come with a particular version of the language installed, along with most common tooling. They are publically available on Docker Hub and save a lot of time in not having to install all of the above features into the image on your own.

Intermediate Images

While the robust images are great, they can be quite memory heavy and ideally we would like to get the best of both worlds:

  • When building the baduk.ninja image we would like to use robust base images. This will allow us to build frontend and backend without losing time for installing languages and tooling.
  • We would like the final image to be constructed based on a minimal Linux setup to save memory

To achieve that we will use three base images:

  • rust:1.73.0 - base image for Rust builds
  • node:14 - base image for Node.js builds
  • frolvlad/alpine-glibc:latest - a minimal Linux Alpine build

The first two will be used as intermediate base images. They will be mounted and used to build backend and frontend respectively. We will grab the optimized executables from the intermediate images and copy them to the lightweight Linux Alpine base image. Afterwards the robust intermediate images will be discarded.

The major steps for our Dockerfile can be summarized as follows:

1) Frontend build

- Mount Node base image

- Copy local, unoptimized frontend source code inside the image

- Install libraries with npm

- Build frontend code

2) Backend build

- Mount Rust base image

- Clone the goban fork directory from Github

- Copy local, unoptimized backend and sgfs code inside the image

- Build backend code

3) Final build

- Mount Linux Alpine base image

- Copy only executables from backend and frontend intermediates

- Install npm and serve tooling (necessary to launch frontend)

- Copy sgfs directory for optimized backend to work properly

4) Add information in the image which ports should be exported

5) Define a default command to launch upon container start

Dockerfile Code

Just as described above, we will start by building our frontend. As you can see below, we start off with picking node:14 as the base image for our intermediate image and alias it as frontend-build for future reference.

# Stage 1: Build the React Frontend
FROM node:14 as frontend-build

Our base image by default has a few basic directories inside. One of them is called app and it is customary to place our app’s code there.

We will use command WORKIDR to set our current working directory to /app/frontend. You can think of this as equivalent of the cd bash command, but WORKDIR will create you a directory in the expected path, if it doesn’t yet exist (whereas cd will not).

# Set the working directory for the frontend
WORKDIR /app/frontend

Once we’ve done that we will set an environment variable (yes, containers have their own set) to let frontend know at which address to communicate with backend. This environment variable is used at frontend build time.

ENV REACT_APP_BACKEND_ADDRESS=baduk.ninja:4000

We copy the frontend code from our local frontend directory into the current working directory in the image (so /app/frontend as specified earlier).

# Copy the frontend source code
COPY ./frontend ./

We then install frontend libraries with npm and build the code. It will create us a build directory with the optimized files that we want to include in the final (non-intermediate) image.

RUN npm install

# Build the frontend in production mode
RUN npm run build

The backend build follows a similar pattern. First we mount a new base image and clone a github repo with the forked goban library we want to use. Then we copy the local backend code and sgfs files into their appropriate paths in the image, so that the directory hierarchy reflects our local project setup.

In the final step we build the backend, which gives us an optimized executable for the final image.

# Stage 2: Build the Rust Backend
FROM rust:1.73.0 as rust-build

# Clone the goban repository
RUN git clone https://github.com/lukaszlew/rust-goban-fork.git /app/rust-goban-fork

# Set the working directory for the backend
WORKDIR /app/baduk.ninja/backend

# Copy the backend source code
COPY ./backend ./
COPY ./sgfs /app/baduk.ninja/sgfs

# Build the Rust backend
RUN cargo build --release

With frontend and backend optimized we move on to creating the final image. We mount Linux Alpine as our base and install some tooling necessary for the optimized frontend to launch in this minimal image.

We also create a baduk.ninja app repository and switch into it. This is where we will copy our optimized code into.

# Stage 3: Create the Production Image
FROM frolvlad/alpine-glibc:latest

# Install Node.js and npm in the final image
RUN apk --no-cache add nodejs npm

# Install serve to serve the frontend
RUN npm install -g serve

WORKDIR /app/baduk.ninja

We reference our previously aliased frontend and backend intermediate images and copy only the optimized code into our Linux Alpine image. Then we also grab the sgfs files necessary for the backend to work properly, as well as run_services.sh file. As mentioned previously, this script, by default, will launch once we start a container based on our image.

# Copy the built frontend from the frontend-build stage
COPY --from=frontend-build /app/frontend/build frontend

# Copy the built Rust binary from the rust-builder stage
COPY --from=rust-build /app/baduk.ninja/backend/target/release/backend .
COPY --from=rust-build /app/baduk.ninja/sgfs ./sgfs

# Copy necessary scripts or files
COPY ./run_services.sh ./

Our .sh file is not executable by default, so we grant it the necessary permissions.

RUN chmod +x /app/baduk.ninja/run_services.sh

We tag our image with information which ports should be exposed.

# Expose necessary ports
EXPOSE 80
EXPOSE 4000

Finally, we define the default launch command for our image - to launch our run_services.sh script. It will invoke both frontend and backend, effectively starting our app.

# Define the command to run your application
CMD ["sh", "/app/baduk.ninja/run_services.sh"]

At this point we would save the Dockerfile and cd into the directory it is located into. Running a docker build command there should start up the creation of the image.

What’s next?

The image we’ve just built works fine, but in a large scale application this setup would be difficult to manage. Large apps tend to work in a global way - you have instances of your app running in many geographic locations. You also want to have control over how many instances of your app are available to the users to meet the demand. Better yet, you would like to launch new containers when the demand increases and shut them off when it decreases to save compute.

Finally, you might want to be able to scale just separate parts of the app - like backend, frontend or other related services. That’s why most large scale apps would not pack both frontend and backend into a single image. A popular approach is to create an app based on microservices, where every major function of the app has their own image, which can be scaled independently. It also makes it easier when deploying updates - you need to update just one micro service, not the whole app.

The microservice approach introduces a question - if my services will be running separately, how will they communicate? That and other issues mentioned earlier can be handled by Kubernetes.

It’s a technology aimed at orchestrating the containers to work together. You can define complex scaling strategies, control the amount of running containers, their placement and communication. Although Kubernetes calls them pods, the basic blocks of the technology are not unlike the containers you’re already familiar with. It’s a very powerful tech to enhance your app with once it grows, so if you’re looking to strengthen your ops skills, Kubernetes is the direction you might want to grow into.