-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pipelines pods FailedMount #3407
Comments
This is likely caused by a non-Docker Kubernetes cluster. |
@Ark-kun I checked that my Kubernetes cluster uses Docker as its container-runtime. I also tried running https://github.com/argoproj/argo/blob/master/examples/artifact-passing.yaml using KFP but i received the same error: Changing the output directory to a directory that does not exist in the base image results in the same error as well. |
Can you try changing the executor back to Docker and submitting that Argo example just using kubectl create -f https://github.com/argoproj/argo/blob/master/examples/artifact-passing.yaml ? If it fails, then it looks like it's an upstream issue that's related to your cluster configuration. What kind of Kubernetes cluster are you using?
If you really mount emptyDir: {} under the artifact file, this error will probably disappear. But it's not a good solution. |
@Ark-kun I got the same original error if i change the executor back to Docker and submitting the Argo example. |
Can you please try to create an issue in the Argo repo? |
@Ark-kun Indeed, when I really mount emptyDir:{} under the artifact file, I was able to get pipelines running. I will try to create an issue in the Argo repo. Thanks for your help! |
Hi Sachau, I am having same problem using PKS kubeflow. Could you plz elaborate how you solved this issue, Appreciate all your help. |
I am getting error like invalid spec: templates.kale-marshal-volume.outputs.parameters.kale-marshal-volume-manifest: k8sapi executor does not support outputs from base image layer. must use emptyDir |
@pvgbabu if you are creating workflow pipelines by applying a yaml file you will have to manually mount an emptyDir (see the docs) If you are creating workflow pipelines using jupyter notebook (e.g. Kubeflow's example simple-pipeline), you will have to add a few lines of code when you are creating the operation arguements to add the emptyDir and mount it to the containter-op. See #1654 (comment) Alternatively, I would suggest that you use PNS executor instead, as that was what I settled on after trying all the available executors. I opened an issue in the Argo repo regarding the mounting of emptyDir and it seems to be not possible currently for k8s api and kubelet executors to support outputs from the base image layer argoproj/argo-workflows#2679 (comment) I was able to find a concised pros vs cons of each executor, hope it will be able to help you decide better on the executor you want to use argoproj/argo-workflows#1256 (comment) (it seems that Argo is using PNS themselves) Hope I was able to help! |
@pvgbabu I have not tried Kale, but I would think that they have a way to add a volume mount? Or you could edit the code generated by Kale? I highly recommend you try the PNS executor first, it would solve the emptyDir issue without the need for you to constantly need to mount emptyDir to every container operation. |
@pvgbabu I just tried Kale, it works with PNS executor, but I had to edit the pipeline yaml's access mode from ReadWriteMany to ReadWriteOnce, since PKS doesn't support ReadWriteMany currently. |
@sachu my KALE working with PNS executor. Thanks for all your help |
I tried the example from https://github.com/kubeflow/examples/tree/master/pipelines/simple-notebook-pipeline to test the functionality of my Kubeflow Pipelines.
I was unable to create the pods for the pipelines.
Below are the event logs from the pods:
Could it be due to the PVC requiring ReadWriteMany access?
I am using VMWare PKS to deploy my Kubernetes cluster.
If my cluster does not support ReadWriteMany access, are there any suggestions on how I can solve this problem?
The text was updated successfully, but these errors were encountered: