-
Notifications
You must be signed in to change notification settings - Fork 185
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Volume Creation failed with - CnsFault error: CNS: Failed to initialize FcdService :Operation timed out #135
Comments
@doctori CNS does respect storage policy specified in the request. CSI driver does send list of shared datastores accessible in the requested zone/region. You can verify this on the vCenter CNS UI. The issue I see is
Can you post VC side log here? To check the logs on the vCenter Server side, search for the logs with the cns prefix at the following locations:
|
this is from the vsanvcmgmtd.log :
so clearly there is something wrong on the vcenter side. |
@doctori Did you get help from vmware support regarding this? Let us know if you need any help. |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
this was an issue from the vcenter |
@doctori: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
@doctori What was the issue with vcenter? I get a slightly different error but might be related
|
@dmc5179 : you might want to check your connectivity between you CSI controller pod and vcenter, looks like the controller can't talk to the vcenter. or you might want to raise the roundTrip value to give the client a chance to get a response after a while (on slow vcenters this helps) |
@doctori I suspect the issue is because we have vCenter on a different port from the default 443. I had to update the OpenShift cluster config when deploying to ensure OpenShift understands how to talk to vcenter on the other port. When I deployed the CSI driver I made sure the configuration included the updated port. It's possible that there is somewhere in the code that doesn't use the config and is locked to 443 or there is another place that I need to configure the CSI driver to use an alternate port for vCenter. I'm going to setup a port forwarding rule on one of our hosts port 443 to route calls to vCenter and see if the CSI driver is happier. If is is then I might have open an issue here about supporting alternate ports. |
@dmc5179 have you tried changing the |
@RaunakShah @doctori We are adding the custom port to the main cloud provider config map which works as the default storage class comes online when we install the cluster. We're also adding the port to the csi-vsphere.conf but that doesn't appear to resolve the issue. I've even gone so far as to create another host as a proxy, forwarding 443 on the proxy to 8443 on vcenter, and installed the cluster against the proxy. When I do that, and leave the port out of any of the configs, the default storage class again works, through the proxy 443 -> vcenter 8443, but the CSI driver fails with the exact same error message. Perhaps there is a service that I need to enable in VMWare somewhere to support the CSI driver? |
@dmc5179 |
@divyenpatel We're reinstalled the cluster several times, including with the port set in the config maps and secrets from the start. Interestingly enough I get a different error if I set the storage policy name:
I get a different error from the PVC:
I'm not sure yet what service I need to turn on, or how, in vCenter to enable to storage policy based management service. |
So I figured out that I cannot use the storage policy in the storage class because I don't have a vSAN. We only have on ESXI host with vCenter. So I'm using the datastore in the storage class. Which results in the error about FCD
How do I turn on the fcd service? |
…ency-openshift-4.19-ose-vmware-vsphere-csi-driver OCPBUGS-45402: Updating ose-vmware-vsphere-csi-driver-container image to be consistent with ART for 4.19
/kind bug
What happened:
In a topology aware environment the CNS creation request doesn't respect the the storage policy and send the full list of datastore available on all the clusters of the vcenter
the FCD service timesout because some datastores are not reachable - some clusters are not in 6.7u3.
What you expected to happen:
the CNS creation request should only holds datastore compatible with the storage policy
How to reproduce it (as minimally and precisely as possible):
Create a storage policy concerning only 2 datastores attached to two clusters with ESXi nodes in 6.7u3 within a environment with multiple other clusters with ESXi 6.7u2 (or even lower)
then create a k8s cluster with this kind of storage class :
Anything else we need to know?:
This situation seems to have appened after a few weeks of runtime (restarting the whole stack didn't fix the behaviour)
Environment:
csi-vsphere version: 1.0.2
vsphere-cloud-controller-manager version: 1.1.0
Kubernetes version: 1.16.4
vSphere version: 6.7.0 build 14368073
OS (e.g. from /etc/os-release): Ubuntu 18.04.2 LTS
Kernel (e.g.
uname -a
): Linux rebrand0026.hosting.cegedim.cloud 4.15.0-50-generic Metadata Syncer for vsphere CSI Driver #54-Ubuntu SMP Mon May 6 18:46:08 UTC 2019 x86_64 x86_64 x86_64 GNU/LinuxInstall tools: Rancher 2.3.4
Others:
The text was updated successfully, but these errors were encountered: