-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Warning message is confusing when pod logs cannot be retrieved #3711
Comments
Thanks for the suggestion! We can only show troubleshooting guide when there is an error, but not a warning. |
happy to fix this |
oh wait is this already done? i.e. just removing 'troubleshooting guide'? |
@jonasdebeukelaer Thanks for offering help! Some helpful information for contribution:
My suggested UX would be to hide the troubleshooting link when the given banner is a warning (it still shows it when the banner is error), but you can take a look and decide if that feels reasonable to you. It already supports hiding the link on ad-hoc usage: https://github.com/kubeflow/pipelines/blob/master/frontend/src/components/Banner.tsx#L72, so we can also dynamically configure it on usages. |
@Bobgy @jonasdebeukelaer I wonder if it is okay to remove the message completely, or at least lower the level to informational (without the exclamation mark and "Warning" prefix). |
/reopen Now it shows fail to retrieve pod logs with error message:
We didn't accommodate for the case when workflow was also missing. |
@Bobgy: Reopened this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
@jonasdebeukelaer do you want to revisit this? |
hey @Bobgy should be a quick one to fix so happy to do it. In what situations can a workflow be missing? |
When user configures a TTL to GC workflows. (We have a default TTL of 1day) In fact, workflow status should be persisted into run details DB rows. So UI shouldn't need to fetch the workflow |
hmm makes sense 👍 |
@Ark-kun this bug: #3711 (comment) must be fixed before logs can be reused from archive. |
@Ark-kun How do you enable the option? |
No update yet, we need someone from community to fix this problem. For us, we are on GCP, so GCP stackdriver auto persists all Kubernetes pod logs. |
@haydnkeung I managed to enable logs persistence. Check the configmap workflow-controller-configmap and see if archiveLogs: true is set. For me it wasn't, even though I'm on KF 1.1, and I had to set it in the config-map.yaml found in your manifest dir under argo/base. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
Dear @ConverJens , did you restart ml-pipeline-ui (kubeflow UX) after editing configmap? |
@ra312 I actually redeployed all ml-pipieline components apart from Minio and Mysql so I don't know which is required. However, I don't think the UI has anythinh to do with this but rather it's the api server that needs restarting. The UI will simply pick up the logs as any other artifact. |
Thanks, @ConverJens! I will try to do the same. |
@ConverJens @ra312 I'm also trying to persist the pipeline pod logs so that if pod gets deleted then logs should be available to the pipeline runs. Is there any other step which needs to be done ? |
@rohitgujral The logs tab will only be populated until the pod is removed. The complete logs will be available as a tar.gz under artifacts instead, called main-logs.tar.gz I think. Note that while the logs tab will always hold the full log, the artifact can lose the final part in the event of your component crashing.I believe that this has to do with the logs not being fully flushed in some instances and in that case, only the logs up the point of error is available. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
closing since the issue has been reported as fixed in #3848 |
/close |
@ra312: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
What steps did you take:
After the pod finishes successfully and is later reclaimed, the following warning is displayed. The message often confuses first-time users, and suggests them to check out the troubleshooting guide.
What happened:
In fact, the logs can be viewed in Stackdriver Kubernetes Monitoring.
What did you expect to happen:
Remove the warning message.
/kind bug
/area frontend
The text was updated successfully, but these errors were encountered: