Docker containers and containerization are popular ways to package and deploy various applications. But, building a Docker image is a skill that begins at the very beginning. From the moment we create a Dockerfile, we should be thinking about the fundamentals of building a Docker image quickly.
In this post, we'll work with an example project to chat through four fundamentals to make building Docker images as fast as possible when using one of the docker commands; the docker build
command.
- Keeping our Docker build context small by excluding things in our source code for a faster image build
- Using a minimal base image for a smaller container image
- Leveraging a multi-stage image build for keeping our final Docker container slimmed down
- How to use BuildKit cache mounts for a fast image build when you use cache but it invalidates
Example project
Let's create a sample Node API project we can work with throughout to build a new docker image. We will leverage Fastify to create an API that we configure via the fastify-cli
.
First, we initialize a project via pnpm
or your favorite package manager. Then we install the fastify-cli
globally via your preferred package manager:
Then, create a new project via the fastify-cli
. We use the --lang=ts
flag to create a TypeScript project:
The details of what the API does aren't that important for this post, so we can stick with what generate
gives us. We can start the API after running pnpm install
via pnpm start
and test it out:
We can test the API out via curl
:
Fundamentals of building a Docker image quickly
Now that we have an example project to containerize, we can start looking at the fundamentals of docker build
to build images fast. First, we need a Dockerfile
and a .dockerignore
file at the root of our git repository.
.dockerignore
file to exclude files from the build context
Fundamental #1: Use a The .dockerignore
file excludes files from the build context during the build process. We don't want to copy over files that aren't needed to run our application or that we generate during the build.
We exclude things like README
files or unnecessary files like our .git
directory. We also exclude artifacts that will get generated during the build, like node_modules
, because we will install those in our build.
Fundamental #2: Use a minimal base image
When building Docker images, it's common to jump to a generic base image like the ubuntu
or node:20
. But, this has downsides that impact the build command and even docker run
down the line.
- Generic base images are often large, which means we have to download a lot of data before we can start building our image layers
- Large base images can make our final image larger than it needs to be because they include a lot of unnecessary dependencies that we don't need
- Large Docker images are slow to build and slow to run
So, instead of reaching for the large base image, we can reach for the slim
version instead. Here is an updated Docker file that uses the slim
version of the node:20
base image:
This smaller base image is 244MB instead of 951MB. A smaller base image will make building the Docker image faster. Why? Because we will have less data to download and package up. Our final image is also smaller because we won't have unnecessary dependencies from the larger base image.
alpine
?
Why not Another option is to use the alpine
base image. But, it's an unofficial Node runtime because it uses musl to implement the C standard library. The difference can cause performance issues, bugs, and application crashes. So for this post, we stick to a slim
image instead.
What's happening in the actual Dockerfile?
Our first RUN
command is corepack enable
so we can access pnpm
without installing it ourselves. Next, we install the fastify-cli
so that we can use it to start our API.
Then we use the COPY
command to copy in our package.json
, pnpm-lock.yaml
, and tsconfig.json
files. With those files copied, we can use another RUN
command to execute pnpm install
to install our dependencies.
Once our dependencies are installed, we can build our TypeScript files and output them into our dist/
folder via the final RUN
command. Lastly, we expose port 3000 for our image to run on and set the CMD
instruction to start our API.
We can build and run the image with the following docker commands from our command line:
Copy only what you need to install dependencies
Notice that we don't copy over our entire repository (i.e., COPY . .
) and only copy the files needed to install our dependencies. If we copy over our entire repository, any code change would invalidate the cache. Invalidating the cache would mean having to reinstall our dependencies. This is a waste of time and resources.
Check out our fast Dockerfiles theory & practice post for more background on how that subtle mistake could cause your docker build
times to explode.
Fundamental #3: Use a multi-stage build
A multi-stage build uses many FROM
statements in our Dockerfile
to create many images. We can copy files from one stage to another, which allows us to create a final image that only contains the files we need to run our application. We can also build stages in parallel, which speeds up our builds.
Here is what our Dockerfile
looks like with a multi-stage build:
We now have four stages in our multi-stage build process:
base
- This is the base image we use to installpnpm
andfastify-cli
dependencies
- We install our dependencies without dev dependencies via the--prod
flagbuild
- We install our dependencies with dev dependencies to build our TypeScript files- Our final image - We copy over our dependencies from the
dependencies
stage and our built TypeScript files from thebuild
stage using thebase
stage as our base image
When we run a docker build
now, we build each stage in parallel. Then in our final stage, we copy the files from the previous build stages into our final container image that the Docker engine can run. The earlier stages get excluded from the final image. The net result is that we have a final image that is 285MB instead of 412MB.
Fundamental #4: Leverage BuildKit cache mounts
BuildKit cache mounts speed up our builds when the cache gets invalidated and a given layer needs rebuilding. The target of a cache mount is persisted across builds, assuming you have a persistent disk to keep them on. See our post on using BuildKit cache mounts in CI for more in-depth details.
Here is our final Dockerfile
with BuildKit cache mounts:
We added a --mount
flag to our RUN
statements that install our dependencies. This tells BuildKit to store the contents of /root/.local/share/pnpm/store
, the virtual store directory for pnpm
, across builds. So the next time we build our image, the virtual store will be mounted, and we won't have to download the complete list of dependencies again.
Conclusion
From the moment we create a Dockerfile, we should be thinking about the fundamentals of building a Docker image as fast as possible.
These fundamentals will help us build images fast. They allow us to leverage as much of our previous build results as possible to avoid unnecessary work. They help us keep our image sizes small and our builds & deployments fast.
40x faster builds with Depot
Our CLI, depot build
, is a drop-in replacement for docker build
that can make your Docker image builds up to 40x faster.
We launch remote Docker image builders supporting x86 and Arm architectures. These remote builders come with 6 CPUs, 32 GB memory, and a persistent 50 GB NVMe cache disk.
With these fundamentals, you can leverage Depot to build your images even faster. We automatically persist your layer cache across builds via a local SSD. It's available instantly during builds and shared with anyone who has access to the project. A developer who builds an image can reuse the result their coworker produced minutes before.
If you want to try out Depot to accelerate your docker build
workflow, sign up and try things out via our quickstart guide.