buildkit: Cannot give a local image to FROM when using docker-container
I have a Dockerfile which references another image, which is stored locally (overrideable with a build arg):
ARG PARENT=ds_test_base:test
FROM ${PARENT}
The image is there:
$ docker image ls ds_test_base:test
REPOSITORY TAG IMAGE ID CREATED SIZE
ds_test_base test eb69be11f1e5 3 hours ago 10.4GB
When I try to build this Dockerfile with --builder xyz, which is backed by a docker-container driver, I get this:
#3 [internal] load metadata for docker.io/library/ds_test_base:test
#3 ERROR: pull access denied, repository does not exist or may require authorization: authorization status: 401: authorization failed
------
> [internal] load metadata for docker.io/library/ds_test_base:test:
------
Dockerfile:2
--------------------
1 | ARG PARENT=ds_test_base:test
2 | >>> FROM ${PARENT}
3 |
4 | MAINTAINER "Openmail"
--------------------
error: failed to solve: ds_test_base:test: pull access denied, repository does not exist or may require authorization: authorization status: 401: authorization failed
If I try to build it with exactly the same parameters, but omitting --builder xyz, it builds just fine.
$ docker version
Client: Docker Engine - Community
Version: 20.10.8
API version: 1.41
Go version: go1.16.6
Git commit: 3967b7d
Built: Fri Jul 30 19:54:27 2021
OS/Arch: linux/amd64
Context: default
Experimental: true
Server: Docker Engine - Community
Engine:
Version: 20.10.8
API version: 1.41 (minimum version 1.12)
Go version: go1.16.6
Git commit: 75249d8
Built: Fri Jul 30 19:52:33 2021
OS/Arch: linux/amd64
Experimental: false
containerd:
Version: 1.4.9
GitCommit: e25210fe30a0a703442421b0f60afac609f950a3
runc:
Version: 1.0.1
GitCommit: v1.0.1-0-g4144b63
docker-init:
Version: 0.19.0
GitCommit: de40ad0
$ docker info
Client:
Context: default
Debug Mode: false
Plugins:
app: Docker App (Docker Inc., v0.9.1-beta3)
buildx: Build with BuildKit (Docker Inc., v0.6.1-docker)
scan: Docker Scan (Docker Inc., v0.8.0)
Server:
Containers: 2
Running: 2
Paused: 0
Stopped: 0
Images: 27
Server Version: 20.10.8
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
userxattr: false
Logging Driver: json-file
Cgroup Driver: cgroupfs
Cgroup Version: 1
Plugins:
Volume: local
Network: bridge host ipvlan macvlan null overlay
Log: awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog
Swarm: inactive
Runtimes: io.containerd.runc.v2 io.containerd.runtime.v1.linux runc
Default Runtime: runc
Init Binary: docker-init
containerd version: e25210fe30a0a703442421b0f60afac609f950a3
runc version: v1.0.1-0-g4144b63
init version: de40ad0
Security Options:
apparmor
seccomp
Profile: default
Kernel Version: 5.4.0-81-generic
Operating System: Ubuntu 20.04.3 LTS
OSType: linux
Architecture: x86_64
CPUs: 8
Total Memory: 30.58GiB
Name: ip-10-150-29-76
ID: ZJFE:DJF7:DUGP:WOGK:A66E:IVQS:X6HB:CKP4:SBRG:VROU:ZIBO:GWXF
Docker Root Dir: /var/lib/docker
Debug Mode: false
Registry: https://index.docker.io/v1/
Labels:
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
WARNING: No swap limit support
$ docker image ls moby/buildkit
REPOSITORY TAG IMAGE ID CREATED SIZE
moby/buildkit buildx-stable-1 2b537a02e2d9 6 weeks ago 144MB
About this issue
- Original URL
- State: open
- Created 3 years ago
- Reactions: 7
- Comments: 47 (29 by maintainers)
Understood. But isn’t the point of buildkit and the various drivers so that we can open new capabilities? buildkit does this phenomenally well, but some of those capabilities are not yet retrofitted into the docker driver. So you have to choose between:
FROMI think we are trying to bridge this gap: if docker driver could do everything buildkit can (which it eventually will, I understand), and buildkit containerized could do everything docker could (again, within reason), then the choice between features Docker and features buildkit (and if you need both, you are out of luck) wouldn’t be a problem.
Sure, and I built tons of images that way in the early days. And when I could move to multistage after it existed, I did.
But is it a fair assumption that every (or even a majority) of builds are simple linear “start at point A, go through some intermediates, get to point Z” and thus multistage candidates?
It is very common to have a base image (A), build some “golden base” (B), then some intermediates (C, D, E, F), then some finals (Q, R, S, T, U, V, W, X, Y, Z). This “tree” isn’t a single build that can go in a multistage Dockerfile, and the intermediates aren’t stored in a remote registry: they might be private, they might be part of a testing process and aren’t valid to be stored until the finals (Q, R, S, T, U, V, W, X, Y, Z) are generated and fully tested (if at all). The local cache holds all of those interim images, and the later steps absolutely are not a single multistage build, simply cannot be. They cannot even be run at the same time (like
bakewith abake.hcl).These are separate build processes, might be running at separate times. This isn’t a single multi-stage Dockerfile (imagine what a nightmare that single file would be), and isn’t run as a single step or stage in a CI process.
Let’s be practical: someone has been doing this for years with
docker build. It works great (it really does, hats off). They see buildkit, love what it can do, need those features that do not yet exist in the docker driver, so they try to go containerized.And their entire build process breaks. Because step 1 stored output B, and step 2 cannot find B for C,D,E,F (repeat for step 3).
That’s just it. The above very common use case describes how “one command” isn’t practical here. It has to be multiple commands. Yet
docker buildcombined withFROMand local caches made those multiple commands dead simple to grasp and use. That was a key part of docker’s adoption.The part I am having a really hard time figuring out is, why the strong objection to proper caching inside buildkit container? You already store blobs, you already have the OCI layout, the number of additional steps to get to parity with default docker behaviour is so small (and people have offered to help build it). Why the resistance?
While explaining this issue to some people, I realized the most basic use case why it matters, and why multi-staged docker files don’t solve it, why bake and build contexts only partially do it.
Let’s say that I am developing an image that normally is public, call it
foo/myimage. Part of my build and test process is that there are downstream images, completely independent, that depend onfoo/myimage, call thembar/imageandbar/other.A normal, sane, build process has me do the following:
foo/myimagedocker build -t foo/myimage:testout .foo/myimage:testout, since I have absolutely no idea if that is the finally version until I work up the downstream dependenciesbar/imageandbar/otherdocker build -t bar/image:somethinganddocker build -t bar/other:elsedocker push foo/myimage:testoutThe above is a great, simple chaining process, very Docker-ish, and 100% depends on
foo/myimage:testoutbeing available in some local read-through cache.Even with build-contexts, how would I do my normal process? What can I give to
docker build -t bar/image:somethingthat will point it at local output? buildkit container-drive only understands a networked registry derived from theFROMimage ref, and build contexts require that I explicitly give it the name (meaning yet another part of my build process I need to edit and manage, in addition to theDockerfile), but it also has no “source” that I can point at the output of the first.In theory, I could do a
docker saveon my first image, and then untar it, or maybe build the output with-oto a local file, but my process got much more complex, and if my testing succeeds, I need to rebuild it again in order to push it. Caching will make it more efficient, but yet another step to run.Am I doing a decent enough job explaining how this breaks normal processes?
As far as I can tell - please do correct me - the buildx bake context stuff allows you to, essentially, “alias” a
FROM(or--from=) in a Dockerfile or other builder to one of a local directory, an image in a registry, or possibly the results of a previous stage (based on this).It looks pretty hesitant to use that target reference:
I don’t understand how this manages to store the output of one to the other, especially with containerized builder. Is it just because it is building both at once, so the builder knows about all of the outputs and can “hold onto” them?
More important UX question: Does this mean that, in order to do the simple “use a a local cached image” that we have been used to for docker since day 1, where I do
docker build -t somelocalimage:foo -f Dockerfile1 .and then consume it in another withFROM somelocalimage:foo, I now need to:bake.hclin addition to myDockerfilesall to replicate a “docker-simple” ™ functionality from the existing flow?
If you switch to the default builder (
docker buildx use default), bothdocker buildx buildanddocker buildshould uses the same storage, in which case you can refer to the image.If you are on a recent version of Docker Desktop or the Docker Engine, you can enable the containerd image store integration. That provides a multi-arch store without requiring a docker-container builder. It’s still in preview, but work is progressing and will be moved out of “experimental” in the not too distant future; see https://docs.docker.com/storage/containerd/
Make sure to switch the builder instance back to the default (not the container-builder, otherwise it continues to use the
docker-containerdriver)You don’t need a new flag. This is already supported.
--build-context myimage:1.2.3=docker-image://image. The source can be a local directory, git repository, docker image, URL to tarball, or another build definition inbakefor chaining. Adding another source that is path to OCI file layout does not conflict with the buildkit design.@felipecrs Bake has a different security model indeed. Atm. for local bake files it is allowed for them to point to any local path. For remote files, we have disabled parent access. In future releases, the plan is to move to https://github.com/docker/buildx/issues/179 to control what is allowed and what is not.
FWIW, this is exactly what we raised in #2210 . builder-container has a regression in functionality compared to docker builder, at least as far as cache is concerned.
I had looked into the containerized builder works, and was surprised that it actually was missing only a few things to make it work. If I recall correctly:
FROMlookup to look in local cache before going to remote registryYou could do other things, like
loadandsaveequivalents, which would be helpful, but not, I believe, critically necessary.I still know of a few people willing to help with it a bit, but it has been 9 months since we opened the issue.
https://github.com/docker/buildx/blob/v0.8.0/docs/reference/buildx_build.md#build-context https://github.com/docker/buildx/blob/v0.8.0/docs/reference/buildx_bake.md#defining-additional-build-contexts-and-linking-targets