-
Notifications
You must be signed in to change notification settings - Fork 123
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Docker image outdated #2646
Comments
I didn't even realize there was a Dockerfile here. I think the problem is that that no contributor seems to use it regularly so it got neglected. Thanks for pointing it out. |
Actually, it seems to have been built 2 weeks ago. This matches with this Docker image, but I'm not sure if it is the actual image that we generate in Github Actions here. Was this the image that you were testing? |
Ok, on closer inspection the Action seems to only build the "latest" tag (which builds off of the We could just regenerate all Docker image tags on all pushes to @gruhn @panasenco Pinging because you two seem to be the ones that have worked the most on the Docker images. |
Haven't used Scryer in a while, but will take a stab at this. There are a couple of things going on here. First, the Dockerfile needs to be updated and tested locally to make sure Scryer still starts. @phmarek , is there a reason to use unstable or testing? Even stable-slim doesn't have any critical vulnerabilities. Then, to @bakaq's point, the base image of at least the latest N tags could be updated on a regular basis using a scheduled job. |
Probably relevant: #2013 |
Well there we go, thanks @bakaq! Looks like we pinned it to be old on purpose 😅 |
Nah, we're good. I just wrote a new version of the Dockerfile that keeps the rust builder and executable versions consistent, and it works now: https://github.com/panasenco/scryer-prolog/blob/master/Dockerfile Next, the CI update... |
@bakaq , I can't find any easy way to do something like "rebuild the latest N tags". Can you think of any approaches? If not, would it be acceptable to keep the existing tags frozen and ask security-conscious folks to use I just don't want to write something crazy that no one but me can maintain. 😅 |
Also occurs to me that even if we found a way to rebuild the previous tags, they're all using |
I'm not very familiar with Docker, so I guess I won't be able to help a lot here. I guess a way would be to find which tags are published in the Docker repository and only update them, because then I think we could just delete tags to deprecate them, but I have no idea how you could do that in a way that is not incredibly obscure.
Probably not, because
I can't think of a way to solve this that doesn't involve making branches for the old releases to fix this. But also, I don't think |
I'm also looking into this now and it is annoying. I guess we could have a scheduled job that
I can imagine that this will break often though. We would have to maintain a Dockerfile that's compatible with the repos states at each version tag.
Don't worry, I'm also committed to maintain at least the docker/CI business 😉 |
I'm starting to lean towards the lazy way out as well :S Maybe the maintenance overhead is not worth it, for a Docker image that's rarely used. I would even suggest to remove all those version-tag based images and only keep the |
How about producing a minimal image, ie. one that only contains the Basically something like this: # Installation as now
RUN mkdir /image
RUN rsync -vaR \
/usr/local/bin/scryer-prolog \
/usr/lib64/... and other files \
/image/
# Last stage, return actual image
FROM scratch
COPY --from=0 /image/ /
CMD ["/usr/local/bin/scryer-prolog", "--no-add-history"] That's what we're doing, see https://gitlab.opencode.de/brz/containerplattform/minimal-image. Another idea would be to remove all unnecessary packages, but that means fighting If there's a list of required files, and from that a list of required packages, a periodic job can check whether one of these changed - and only then run the GitHub build process. |
Oh, sounds like this got auto-closed when #2647 was merged. @mthom can you re-open?
I agree that's a good idea in general. But we would still need some process to update the base image of old versions. It wouldn't be that much work to do it manually once in a while, but @mthom owns the DockerHub credentials. So, either we have to ask him to do it for us, or he needs to give one of us permissions. I think both is not a fair ask. So it should be some automatic process. But I can't think of a way to make this run trivially stable long-term. For example, say we go with the approach I suggested:
Then at some point a dependency is updated in the Dockerfile, which is incompatible with version 0.9.2 or whatever. Then one month later when the scheduled pipeline runs, the image build fails. This stuff is just annoying to fix. It's not just: find problem -> open PR -> merge -> done. Maybe we have to create a branch off-of the the version tag and adjust the Dockerfile there. Maybe we have to re-push the Git version tag, to be align it with that change (have ask @mthom do that for us). Then we probably also have to ask @mthom to manually re-push the docker image. That's all not a big deal. But the size of the deal should be proportional to the value, that the Docker image provides. And I suspect not many people care.
But where do we maintain this list of files? Usually you would just spell it out in the Dockerfile. When something changes, we change the Dockerfile. But where do we maintain this information for past versions of Scryer. We could have a dedicated branch for each version as @bakaq suggested. But now we are starting to bend the repo just to accommodate the Docker image. If there is a simple process to just keep the base image of the last release up to date, it's probably not much more work to keep the base image of all prior releases up to date. I'm thinking: some tool like Dependabot but I haven't found anything that really fits the use case. |
Please use Debian Testing or Unstable when building the Docker image, and/or refresh them more often.
Thanks!
The text was updated successfully, but these errors were encountered: