How can you improve the state of your infrastructure automation?

➡️ Take the Self-Assessment

Docker

Docker Image Layers – What They Are & How They Work

docker image layers

Subscribe to our Newsletter

Mission Infrastructure newsletter is a monthly digest of the latest posts from our blog, curated to give you the insights you need to advance your infrastructure game.

Docker containers are isolated environments created from a filesystem template called an image. Images store data as a stack of multiple layers, a design that improves efficiency by allowing layers to be cached and shared between different images.

In this article, we’ll explain image layers and how they benefit container operations. We’ll show how to inspect an image’s layers and share some best practices for optimizing your own builds.

What we will cover:

  1. What is a Docker image layer?
  2. How do Docker image layers work?
  3. Benefits of Docker image layers
  4. Which Dockerfile instructions create a new image layer?
  5. How to inspect a Docker image layers
  6. How to optimize Docker image layers
  7. Docker layers best practices

What is a Docker image layer?

A Docker image layer is a set of filesystem changes applied on top of the previous layer. Individual layers include just the files that a specific Dockerfile instruction added, changed, or removed. Each instruction starts a new layer, although some of these layers may be empty.

For example, the following Dockerfile creates two new layers, one for each of the RUN instructions:

FROM alpine:latest

RUN apk update &&\
	apk add nodejs

RUN echo "console.log('Hello World')" > main.js

The apk update and apk add nodejs commands are first run as a new layer on top of the last layer in the alpine:latest base image. This layer will include only the new and changed files after Node is installed in the filesystem.

Next, a new file called main.js is created in a second layer. This layer will only include main.js — it will not include the full filesystem from the previous layer. If you modify files that already existed in the previous layer, then the affected files will be copied to the new layer, but others will not.

When you build this image, you’ll see Docker execute the two RUN instructions and export the new image’s layers. The final image is created by assembling the layers into a stack so the topmost layer (main.js) sits on top of the layers below it.

$ docker build -t demo-image .
[+] Building 6.1s (8/8) FINISHED                                                                                             docker:default
 => [internal] load build definition from Dockerfile                                                                                   0.1s
 => => transferring dockerfile: 179B                                                                                                   0.0s
 => [internal] load metadata for docker.io/library/alpine:latest                                                                       1.5s
 => [auth] library/alpine:pull token for registry-1.docker.io                                                                          0.0s
 => [internal] load .dockerignore                                                                                                      0.1s
 => => transferring context: 2B                                                                                                        0.0s
 => [1/3] FROM docker.io/library/alpine:latest@sha256:21dc6063fd678b478f57c0e13f47560d0ea4eeba26dfc947b2a4f81f686b9f45                 0.6s
 => => resolve docker.io/library/alpine:latest@sha256:21dc6063fd678b478f57c0e13f47560d0ea4eeba26dfc947b2a4f81f686b9f45                 0.1s
 => => sha256:21dc6063fd678b478f57c0e13f47560d0ea4eeba26dfc947b2a4f81f686b9f45 9.22kB / 9.22kB                                         0.0s
 => => sha256:2c43f33bd1502ec7818bce9eea60e062d04eeadc4aa31cad9dabecb1e48b647b 1.02kB / 1.02kB                                         0.0s
 => => sha256:4048db5d36726e313ab8f7ffccf2362a34cba69e4cdd49119713483a68641fce 581B / 581B                                             0.0s
 => => sha256:38a8310d387e375e0ec6fabe047a9149e8eb214073db9f461fee6251fd936a75 3.64MB / 3.64MB                                         0.4s
 => => extracting sha256:38a8310d387e375e0ec6fabe047a9149e8eb214073db9f461fee6251fd936a75                                              0.1s
 => [2/3] RUN apk update && apk add nodejs                                                                                             2.7s
 => [3/3] RUN echo "console.log('Hello World')" > main.js                                                                              0.4s 
 => exporting to image                                                                                                                 0.5s 
 => => exporting layers                                                                                                                0.4s 
 => => writing image sha256:c465ff41423ff96e212217d7b7fe613295f01526bab26099fcf32d06566233c6                                           0.0s 
 => => naming to docker.io/library/demo-image                                                                                          0.0

Each layer is essentially a diff of the changes between the current filesystem state and the previous layer. An image is a collection of one or more layers that together define a container’s initial filesystem content.

How do Docker image layers work?

Docker image layers implement a union filesystem. Each layer includes only the changes its build stage made, but unionization combines an image’s layers to produce a single logical filesystem. This occurs transparently during container runtime.

When you create a container, Docker reads the layers within the image and merges them into a union filesystem for the container to use. No content is copied between layers — the layers are simply overlaid by the union process. 

Docker supports several different unionization systems; overlay2 is the default, but alternatives, including BTRFS, VFS, and ZFS, can be configured instead.

Docker image layers and immutability

Docker image layers are always immutable. Once they’re created, they’re read-only and cannot be modified. Any changes, such as adding or removing files, must be applied in a new layer.

Immutability raises a problem when starting containers. The container may need to write new content to its filesystem, such as temporary files created by the application. 

To solve this issue, Docker adds a transient read-write layer to the top of the union filesystem’s layer stack. This allows the container to write into the filesystem seamlessly, but changes are lost when the container is stopped.

Benefits of Docker image layers

Docker’s layering system makes image operations more efficient and performant. Layers contain just the files that were actually changed by a Dockerfile instruction, so they can be shared between multiple images.

  • Efficient reuse – Shared layers between images are cached and not duplicated, reducing storage needs.
  • Faster deployments – Unchanged layers are reused, which speeds up builds and reduces push/pull time.
  • Incremental updates – Only modified layers are transferred, minimizing bandwidth usage.
  • Build consistency – Each layer maps to a specific Dockerfile instruction, making builds reproducible.
  • Simplified troubleshooting – The layered structure helps identify which step introduced an issue.
  • Modularity and portability – Layers help keep things modular, making images easier to move between environments. The same image runs the same way everywhere.

Sharing layers between images

Here’s a Dockerfile similar to the one shown above. The first RUN instruction hasn’t changed, but the second instruction now writes to a different file:

FROM alpine:latest

RUN apk update &&\
	apk add nodejs

RUN echo "console.log('Second Image')" > demo.js

When you build this image, Docker recognizes that the first layer is already cached on your machine. It was created when the earlier image was built. As a result, the build process for the second image is much quicker because only the second RUN instruction needs to be executed.

$ docker build -t second-image .
[+] Building 0.9s (7/7) FINISHED                                                                                             docker:default
 => [internal] load build definition from Dockerfile                                                                                   0.0s
 => => transferring dockerfile: 181B                                                                                                   0.0s
 => [internal] load metadata for docker.io/library/alpine:latest                                                                       0.4s
 => [internal] load .dockerignore                                                                                                      0.0s
 => => transferring context: 2B                                                                                                        0.0s
 => [1/3] FROM docker.io/library/alpine:latest@sha256:21dc6063fd678b478f57c0e13f47560d0ea4eeba26dfc947b2a4f81f686b9f45                 0.0s
 => CACHED [2/3] RUN apk update && apk add nodejs                                                                                      0.0s
 => [3/3] RUN echo "console.log('Second Image')" > demo.js                                                                             0.2s
 => exporting to image                                                                                                                 0.1s
 => => exporting layers                                                                                                                0.1s
 => => writing image sha256:3e03e8dcc6edb251679dcc9afa7ecf22738ab3b9059bd881912e8f59b5b2f7d2                                           0.0s
 => => naming to docker.io/library/second-image                                                                                   0.0s

The command’s output shows CACHED next to the RUN instruction that installs Node, demonstrating the previously built layer is being reused. The build completes in just 0.9 seconds as a result.

Preventing storage and network resource wastage

The ability to reuse layers significantly reduces storage and bandwidth requirements. For instance, you can adapt the example above to create several images that install Node on top of the alpine base image, but those layers will only need to be stored once on your machine.

Similarly, image registries only need to store the new layers they don’t already have. When users pull an image from a registry, they download only the layers that are missing from their local cache. You can demonstrate this by first pulling an image like node:18:

$ docker pull node:18
18: Pulling from library/node
fdf894e782a2: Pull complete 
5bd71677db44: Pull complete 
551df7f94f9c: Pull complete 
ce82e98d553d: Pull complete 
6399a464889d: Pull complete 
a3c94c84d15d: Pull complete 
2cd8c50fd8ca: Pull complete 
247468edfd9a: Pull complete 
Digest: sha256:b57ae84fe7880a23b389f8260d726b784010ed470c2ee26d4e2cbdb955d25b12
Status: Downloaded newer image for node:18
docker.io/library/node:18

You can see that Docker had to pull all the layers because none of them had been previously pulled or built. But if you then pull node:20, Docker can reuse the layers that are shared with node:18:

$ docker pull node:20
20: Pulling from library/node
fdf894e782a2: Already exists 
5bd71677db44: Already exists 
551df7f94f9c: Already exists 
ce82e98d553d: Already exists 
28c7f8675398: Pull complete 
cd36ddf19b49: Pull complete 
24f32f277913: Pull complete 
5dd152761f34: Pull complete 
Digest: sha256:f4755c9039bdeec5c736b2e0dd5b47700d6393b65688b9e9f807ec12f54a8690
Status: Downloaded newer image for node:20
docker.io/library/node:20

The first four layers have the same IDs (content hashes) as layers also included in the node:18 image. Docker shows Already exists for those layers because they were previously downloaded when pulling node:18.

Debugging builds using intermediate layers

Containers normally start from complete images, but these are simply layers that have been assigned a tag. It’s technically possible to start a container from any layer using its ID (docker run <layer_id>), including intermediate layers produced by Dockerfile build instructions.

Starting a container using an intermediate build layer can be a helpful way to debug build problems. This technique was commonly used in older Docker releases as intermediate layers were preserved by default. 

The IDs associated with each layer were displayed in the build output. Unfortunately, the switch to BuildKit in recent Docker versions prevents the convenient use of intermediate layers because they’re no longer stored after the build completes.

Which Dockerfile instructions create a new image layer?

Not every Dockerfile instruction creates a new filesystem layer. Some instructions only add image metadata, which doesn’t affect the image’s filesystem content. These instructions will still add a layer to the stack, but it will be empty and show a size of 0B.

The RUN, COPY, and ADD Docker instructions modify files and create new filesystem layers. All other instructions, including LABEL, CMD, ENTRYPOINT, ENV, and EXPOSE, only set image metadata so they create empty layers.

How to inspect a Docker image layers

The docker history command allows you to view all the layers within a particular image:

$ docker history demo-image
IMAGE          CREATED          CREATED BY                                      SIZE      COMMENT
aaf4e78a28b8   5 seconds ago    RUN /bin/sh -c echo "console.log('Hello Worl…   27B       buildkit.dockerfile.v0
<missing>      44 minutes ago   RUN /bin/sh -c apk update && apk add nodejs …   66.7MB    buildkit.dockerfile.v0
<missing>      12 days ago      CMD ["/bin/sh"]                                 0B        buildkit.dockerfile.v0
<missing>      12 days ago      ADD alpine-minirootfs-3.21.0-x86_64.tar.gz /…   7.84MB    buildkit.dockerfile.v0

Here you can see that the layers in our demo image match the instructions from our Dockerfile. The last two layers are inherited from the base image referenced in the Dockerfile’s FROM instruction.

The command displays the ID of each layer, its creation time and size, and the Dockerfile instruction that created it. The IDs of intermediate layers show as <missing> because the image was built using BuildKit. As explained above, BuildKit does not store intermediate image data, but older Docker versions would have kept these layer IDs intact.

You can also inspect the content of image layers using Dive. Dive is a popular open-source community tool for visualizing container images. It provides an interactive terminal interface.

Download and install Dive from its GitHub Releases page. You can then run dive <image> to view the layers in an image:

$ dive demo-image
docker image layers example

The interface displays the image’s layers in the top-left. You can switch between layers using the arrow keys. The details of the selected layer are shown below the list, while the file tree on the right lets you inspect the layer’s filesystem.

You can learn more about using Dive in the documentation.

How to optimize Docker image layers

Docker’s image layers and union filesystem make container operations efficient, but there are still steps you can take to keep your builds fast and small. The following three best practices will help you optimize your layer structures to reduce waste.

1. Combine multiple commands as one Dockerfile RUN instruction

Chaining multiple commands into one RUN instruction is frequently more efficient than splitting them across separate instructions. This results in one image layer being created instead of a new layer for each command. As a result, files will only be copied once, even if they’re used by multiple commands.

Try to change the following…

RUN apt-get update
RUN apt-get install -y nodejs
RUN apt-get install -y libjpeg-dev

…to this alternative:

RUN apt-get update &&\
	apt-get install -y nodejs libjpeg-dev

A single image layer will now contain all the filesystem changes.

2. Run instructions that rarely need rebuilds first

Efficient use of the Docker build cache depends on Dockerfile instructions appearing before the ones that have frequent changes because they are less likely to need a rebuild., The build cache is invalidated as soon as one layer changes, so all the following layers must then be rebuilt — even if their own content remains the same.

Take the following Dockerfile:

COPY src/ src
COPY package.json package.json
RUN npm install

Any changes to the code in the src directory will invalidate the layer cache for the following instructions. This means the npm install instruction will need to be repeated whenever the code changes, wastefully creating new layers. Build time will increase, and more disk space will be used.

You can fix the issue by rearranging the instructions so src is copied after the npm install command. Source code changes will then be layered on top of the npm instruction’s previously cached filesystem.

COPY package.json package.json
RUN npm install
COPY src/ src

3. Consider squashing images into a single layer

Images output a stack of layers by default. As we’ve seen in this guide, this enables layers to be reused when building other images or pushing and pulling from registries. But sometimes, squashing your images down to a single layer is preferable.

Squashing can reduce image size, particularly when your build process causes the same file to be written to multiple times. Those files would normally appear in each layer that changes them, even though only the version in the last layer is actually accessible in the image.

The docker build command’s --squash flag can squash images when Docker’s legacy image builder is enabled. At the end of the build, Docker unionizes the contents of all the layers created and stores the resulting filesystem as a single new layer. This layer becomes the final image.

Unfortunately, squashing with the --squash flag has been deprecated and removed from BuildKit. If you’re using BuildKit — the default builder in recent Docker versions — using --squash will display a warning, and your layers will not be squashed:

WARNING: experimental flag squash is removed with BuildKit. You should squash inside build using a multi-stage Dockerfile for efficiency.

It’s possible to achieve a similar effect without --squash by writing a multi-stage Dockerfile. In the last stage, you should copy in solely the content from the previous stages that you want your final image to include:

FROM node:18 AS node
COPY package.json .
RUN npm install
RUN npm build

FROM alpine:latest
COPY --from node /demo/output/path /bin/app
CMD ["/bin/app"]

The output image will consist of alpine:latest with the /bin/app path layered on top. The package.json file and the results of npm install and npm build won’t be present in the filesystem, reducing the image’s size.

We also encourage you to explore the ways Spacelift offers full flexibility when it comes to customizing your workflow. You can bring your own Docker image and use it as a runner to speed up deployments that leverage third-party tools. Spacelift’s official runner image can be found here.

 

If you want to learn more about what you can do with Spacelift, check out this article, create a free account today, or book a demo with one of our engineers.

Docker layers best practices

Best practices for Docker layers focus on improving build speed, reducing image size, and ensuring cache efficiency. Here are the tips to keep in mind when working with Docker image layers:

  • Order instructions by volatility – Place frequently changing instructions (e.g., COPY . ., RUN npm install) after static ones to optimize caching.
  • Minimize layers – Combine related commands using && in a single RUN to reduce image layers without sacrificing readability.
  • Use multi-stage builds Compile or build in one stage and copy only the necessary artifacts to the final image to reduce size and avoid exposing sensitive data.
  • Leverage .dockerignore – Exclude unnecessary files from the build context to avoid invalidating the cache and bloating the image.
  • Clean up in the same layer – Remove temp files or package managers’ caches within the same RUN command to keep the final layer small.
  • Use official or minimal base images – Choose slim images (like alpine or debian-slim) unless full-featured ones are necessary.

Key points

Docker images consist of layers that are merged into a union filesystem at runtime. Each layer contains just its own filesystem changes relative to the previous layer in the stack. This improves image efficiency as layers can be cached and reused by other builds.

The layering system is designed to be transparent to containers and developers. You don’t need to know exactly how it works to create and use Docker images. However, understanding layers in more depth allows you to optimize your images for faster builds and smaller sizes.

Ready to learn more about Docker images? Check out Spacelift’s guide to writing a Dockerfile or explore how to harden your images in our Docker security best practices post.

Solve your infrastructure challenges

Spacelift is a flexible orchestration solution for IaC development. It delivers enhanced collaboration, automation, and controls to simplify and accelerate the provisioning of cloud-based infrastructures.

Learn more

The Practitioner’s Guide to Scaling Infrastructure as Code

Transform your IaC management to scale

securely, efficiently, and productively

into the future.

ebook global banner
Share your data and download the guide