-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Elastic Agent docker: new revisions are not getting released #24198
Comments
I'm not sure if the goal should be to introduce a separate build process. It is unfortunate that the build is broken and we should ensure on the Beats / Agent end, we detect such problems before they are merged. Is the package development really blocked? It seems the master snapshots are available so development against mater should still work? |
It is, because we can't test against current/older releases. Let me bring few issues: elastic/integrations#736 - I can't reverify if the problem is fixed for 7.11. Also:
My impression is that there is too much coupling in the build process, that even correct products/bugfixes are blocked from releasing by a single item. |
Pinging @elastic/agent (Team:Agent) |
This issue doesn't have a |
Why are old versions a problem? |
It's because of affected branches (backport not released):
7.11 is old enough that we'll skip testing for some packages (compatible with newer Kibana versions) EDIT: Yet another build has just failed, which probably means next 12h of delay. |
@mtojek Just to clarify, the problem is because the unified process produced artifacts are olds and do not include the latest fixes, the new build arent completing? First, are we responsible for the failure of the build? Are our beats or Elastic Agent breaking the build? This we can control and prioritize, if this is not the case we need to look with infra, they are looking into better way to notify us. Like @ruflin said, I don't think introducing a new build is the solution. We have many moving parts, Elastic-Agent, Endpoint, Beats, ES, and Kibana, there are a lot of dependencies and things that need to happen to have confidence in the binary. |
Yes.
I think we should put more emphasis on ownership here and be more proactive in emergency situations, not just sit and wait until the next build appears. In this particular situation the image has been built correctly, but it was discovered later that it doesn't boot up correctly. I had an interesting conversation around this issue with @ycombinator . Of course we can improve the test coverage, but there might be always a situation in which this is insufficient. Such cases should be treated differently to reduce the blast and don't block next customers (e.g. integrations developers, etc.). What do you think about introducing a tagging based solution for Docker images? Let's say that the "unified" builder tags the latest built image with the |
I expect future community developers to develop against stable version of the stack, so I would assume this should not happen. A bit similar on our end, I expect us to be more and more able to develop against stable / released versions. Broken / failed builds will keep happening from time to time, be it that we are in control or someone else. One thing that was great during the development of the package-registry was that each PR and each commit to master had its own docker image / tag. So in case of a broken "latest/SNAPSHOT", a temporary tag could be used. It would be nice, if we would have something similar for the SNAPSHOT builds, so not only having 8.0.0-SNAPSHOT but also 8.0.0-3ac34-SNAPSHOT available so we could go back a few days in case things are broken. AFAIK this does not exist yet? This is very similar to what @mtojek proposed or would be a requirement because otherwise a |
we have short of that, recently we have added the package to the main pipeline, so every time you merge in the master and the build reach the elastic-agent package stage, if that stage end well, a bunch of Docker images would be published in our repository.
ARM Docker images are also published. PRs also publish the Docker images
|
Are these images publicly available ( |
you need to login to access that namespace, we can publish them in another public place |
Yes, that would be a good idea. I don't know green/red stats for packaging of Elastic Agent, but I hope it's relatively rare, so we can leverage from such tags. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
Hi,
due to the bug introduced to the 7.x (and 7.12) branch, the latest snapshot is failing. The bug has been fixed in #24163 and #24161 , but the Docker image hasn't been released due to other issues in the unified release process.
With this problem in SNAPSHOT the package development is blocked for few days in elastic/integration, elastic/elastic-package (all statuses are red). We can hardcode the correct Docker image in both repos, but it means for us a lot of operational work - start a day by reviewing all Beats branches and check if there might be a blocker (failed build), then update a map of hardcoded image references in elastic-package (for 7.13, 7.12, 7.11).
The goal of this issue is to figure out and implement a way of publishing Agent's images independently from other potentially problematic parties.
cc @andresrc @ph @ruflin @ycombinator
The text was updated successfully, but these errors were encountered: