-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pipelines upstream connect error or disconnect/reset before headers. reset reason: connection failure #4469
Comments
I have the same issue with Kubeflow 1.1. The readiness probe is failing on the ml-pipeline-ui pod in the user namespace.
|
This is the only thing that stands out to me. Here is my Kubernetes version: |
I have the same issue after installing kubeflow 1.1 on kubernetes v1.18. besides, i also encountered an installation issue. |
Similar issue, with a hack to resolve it, discussed here |
Hi all, the solution to this is simple, just make sure your PLEASE NOTE: you will need to recreate all resources in apiVersion: v1
kind: Namespace
metadata:
name: kubeflow
labels:
control-plane: kubeflow
istio-injection: enabled |
Can you give your output for |
Hi @amybachir . I'm facing the same issue with ml-pipeline pod. The readiness probe is failing with 503. Did you find any fix or hack? |
@Bobgy I am also facing the same issue. Given below is the output that you requested. What I did.
but after the above steps I still get the upstream connect error |
Same here (HTTP 503) on prem with version 1.2 with dex auth. namespace has correct labels, and disabling mTLS on ml-pipeline leads to "RBAC: access denied" error. |
So after going through few other issues and using their 'hack' it worked (for Azure KF1.2 deployment) by changing the ISTIO_MUTUAL to DISABLE @berndverst, after comparison of AKS KF deployment with GKE KF deployment. we found that AKS ns/kubeflow was missing sidecars see the above image too. |
So instead of disabling mTLS, suggestion of @jonasdebeukelaer in kubeflow/kubeflow#5561 (comment) solved it for me. |
@solarist I Just checked in my azure deployment but I have the meshPolicy as suggested by @jonasdebeukelaer but still the access doesn't work without disabling the mTLS |
@Junaid-Ahmed94 Sorry you've had to troubleshoot this. I sort of inherited whatever was done for v1.1 and moved it over to v1.2 so no surprise the issue still exists here. For Kubeflow v1.3 the goal is to support modern versions of Istio (I believe that work is on track for the v1.3 release). Hopefully I can remove the bundled Istio then and have everything just work. I don't know too much about Istio configuration. |
UPDATE:The problem is indeed that the istio pods are not being injected to the ns kubeflow, the reason is explained in my comment here: |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
This issue has been automatically closed because it has not had recent activity. Please comment "/reopen" to reopen it. |
/kind bug
I have deployed Kubeflow in Azure using kfctl_k8s_istio.v1.1.0.yaml from the master after having modified the repos section
repos:
name: manifests
uri: https://github.com/kubeflow/manifests/archive/master.tar.gz
#uri: https://github.com/kubeflow/manifests/archive/v1.1-branch.tar.gz
version: master
#version: v1.1-branch
I used port forward as follow
kubectl port-forward svc/istio-ingressgateway -n istio-system 8080:80
Everything looks like working from the UI but pipelines where I get the following error once clicking on the left bar Pipelines menu:
upstream connect error or disconnect/reset before headers. reset reason: connection failure
I do expect that the pipelines UI is shown
Environment:
Kubeflow version as above stated is kfctl_k8s_istio.v1.1.0.yaml
Kubeflow version: (version number can be found at the bottom left corner of the Kubeflow dashboard):
kfctl version: kfctl v1.1.0-0-g9a3621e
Kubernetes platform: Azure AKS
Kubernetes version: v1.16.13
The text was updated successfully, but these errors were encountered: