Docker has sharply risen in popularity in the past years. It has been one of the tools that have changed the way we work as software engineers and DevOps Engineers. From Docker v 17.05 multi-stage build was introduced which helped abandon the older builder pattern with use of stages and target. This post discussed how you can exploit docker multi-stage build to build optimal images suited for dev/test and production with a NodeJs example application.
- You are aware of docker and know the basic docker commands like build, exec
- You know about docker-compose (not a necessity)
Docker multi-stage builds intro
Docker multi-stage build lets us build docker images in stages with multiple FROM statements. Files can be copied from one stage to another. A very good example would be how a 294 MB Golang 1.13 official image (123 MB even with Alpine) can be just as big as the go executable of your application. As Golang is compiled and gives out an executable binary, the first stage can be compiling it and the second stage can be an alpine image (5 MB) just to run that executable. So, if you go app binary is 10 MB your image can be 15 Mb (10 MB binary + 5 MB alpine) rather than the heavy 294 MB official go image or 123 MB alpine go image. You can have a look at an example too.
Below is the official information about docker multi-stage builds:
With multi-stage builds, you use multiple FROM statements in your Dockerfile. Each FROM instruction can use a different base, and each of them begins a new stage of the build. You can selectively copy artifacts from one stage to another, leaving behind everything you don’t want in the final image.
Unfortunately, all the language don’t compile to an executable binary as golang does, still, you can leverage multi-stage builds to craft docker images that serve the purpose better. We look into how to do this below with an open-source node js application example.
Issues before multi-stage build
We are going to see an example Node Js app which is a currency converter API built with Express. Currently, the problems with the Dockerfile and build are as follows:
Nodemon is installed on productionCurrent docker image does not have dev dependencies (runs
)The image size can be made smaller (even though its using alpine)
npm install --production
Following are the current Dockerfile and docker-compose.yml for local development:
FROM node:12-alpine WORKDIR /src COPY package.json package-lock.json /src/ RUN npm install --production COPY . /src EXPOSE 8080 RUN npm config set unsafe-perm true RUN npm install -g nodemon CMD ["node", "index.js"]
As we can see nodemon is installed even on production which is unnecessary on production. Another issue is there are no dev dependencies so tests can’t be run inside docker.
Docker Compose file
web: build: . volumes: - .:/src command: npm start ports: - "8080:8080" environment: NODE_ENV: dev VIRTUAL_HOST: 'currency.test' VIRTUAL_PORT: 8080
Don’t be concerned about the
that is for nginx proxy.
Current image size
Let’s look at how big is this image we got from running
docker build . -t currency-api-original
So currently it is 165 MB, hopefully, we can decrease its size too in this process.
Solution with multi-stage build
Now as we want to have dev dependencies and nodemon on dev builds and only production npm dependencies on production build, the docker related files have been modified as follows:
Dockerfile with multi-stage build
FROM node:12-alpine as base WORKDIR /src COPY package.json package-lock.json /src/ COPY . /src EXPOSE 8080 FROM base as production ENV NODE_ENV=production RUN npm install --production CMD ["node", "index.js"] FROM base as dev ENV NODE_ENV=development RUN npm config set unsafe-perm true && npm install -g nodemon RUN npm install CMD ["npm", "start"]
Let’s analyze what changed here and why? Following are the highlights:
- We start with a base image that has node, then copy needed files to the image like 1-5
- For production we set the
to production and install non-dev dependencies, also notice that we run node (not nodemon)
- Later the last 6 lines of the Dockefile, we create the image from the base and set
to development, then we install nodemon as we want to watch the files on dev
- On dev image build we install all npm dependencies including dev ones so that we can run tests
The builds are more streamlined and we have optimized our docker images to be more environment-specific. We solved the above-mentioned issues and don’t have nodemon and dev dependencies on production and we can run our tests on dev/test. That’s a win!
Docker-compose file after multi-stage build
version: '3.5' services: web: build: context: ./ target: dev volumes: - .:/src command: npm start ports: - "8080:8080" environment: NODE_ENV: dev VIRTUAL_HOST: 'currency.test' VIRTUAL_PORT: 8080
The main change for the docker-compose file is the target:dev in the build parameters.
All the changes made can be viewed in this pull request too. Let’s look at how big is the image now:
We ran the following commands to build the dev and the production images:
- docker build . -t currency-api-dev –target=dev
- docker build . -t currency-api-prod –target=production
We have shaved off ~25 MB from the older image for a production build. It happened because we removed nodemon and some dev dependencies for production. Even for the dev build it is ~21 MB smaller.
Conclusion / tl;dr
The main point here is to build docker images apt for the environment and multi-stage builds are an answer to this problem. You can use the same concept to build images for PHP with composer. For example, the dev build can have xdebug for debugging and production build can have
enabled by default.
Anytime you need different things for different environments, opt for docker multi-stage builds and avoid having 3 different dockerfiles.