On Thu, Feb 23, 2023 at 11:21:53AM -0300, Fabiano Rosas wrote: > I'm not sure if this was discussed previously, but I noticed we're not > pulling the images we push to the registry at every pipeline run. > > I would expect we don't actually need to rebuild container images at > _every_ pipeline run, so I propose we add a "docker pull" to the > container templates. We already have that for the docker-edk2|opensbi > images. > > Some containers can take a long time to build (14 mins) and pulling > the image first without building can cut the time to about 3 > mins. With this we can save almost 2h of cumulative CI time per > pipeline run:
The docker.py script that we're invoking is already pulling the image itself eg to pick a random recent job: https://gitlab.com/qemu-project/qemu/-/jobs/3806090058 We can see $ ./tests/docker/docker.py --engine docker build -t "qemu/$NAME" -f "tests/docker/dockerfiles/$NAME.docker" -r $CI_REGISTRY/qemu-project/qemu 03:54 Using default tag: latest latest: Pulling from qemu-project/qemu/qemu/debian-arm64-cross bb263680fed1: Pulling fs layer ...snip... none the less it still went ahead and rebuilt the image from scratch so something is going wrong here. I don't know why your change adding an extra 'docker pull' would have any effect, given we're already pulling, so I wonder if that's just coincidental apparent change due to the initial state of your fork's container registery. Whenever I look at this I end up wishing out docker.py didn't exist and that we could just directly do - docker pull "$TAG" - docker build --cache-from "$TAG" --tag "$TAG" -f "tests/docker/$NAME.docker" as that sould be sufficient to build the image with caching. > We would need to devise a mechanism (not included here) to force the > re-build of the container images when needed, perhaps an environment > variable or even a whole new "container build" stage before the > "container" stage. > > What do you think? We definitely want the rebuild to be cached. So whatever is broken in that regard needs fixing, as this used to work AFAIK. Ideally we would skip the container stage entirely for any pull request that did NOT include changes to the dockerfile. The problem is that the way we're using gitlab doesn't let that work well. We need to setup rules based on filepath. Such rules are totally unreliable for push events in practice, because they only evaluate the delta between what you just pushed and what was already available on the server. This does not match the content of the pull request, it might be just a subset. If we had subsystem maintainers opening a merge request for their submission, then we could reliably write rules based on what files are changed by the pull request, and entirely skip the containers stage most of the time, which would be an even bigger saving. With regards, Daniel -- |: https://berrange.com -o- https://www.flickr.com/photos/dberrange :| |: https://libvirt.org -o- https://fstop138.berrange.com :| |: https://entangle-photo.org -o- https://www.instagram.com/dberrange :|