diff --git a/resources/helm/conda-store/values.yaml b/resources/helm/conda-store/values.yaml index 04ea06a2a..1b3812698 100644 --- a/resources/helm/conda-store/values.yaml +++ b/resources/helm/conda-store/values.yaml @@ -1,14 +1,20 @@ gateway: + # Number of instances of the conda-store-server to run replicas: 1 + # Annotations to apply to the conda-store-server pods annotations: {} + # Resource requests/limits for the conda-store-server pod resources: {} + # Path prefix to serve conda-store-server api requests under prefix: / + # The conda-store-server log level loglevel: INFO + # The image to use for the conda-store-server pod image: name: quansight/conda-store-server tag: "set-by-chartpress" @@ -16,16 +22,20 @@ gateway: imagePullSecrets: [] + # Configuration for the conda-store-server service: annotations: {} auth: + # The auth type to use. One of {simple, kerberos, jupyterhub, custom}. type: simple simple: + # A shared password to use for all users. password: kerberos: + # Path to the HTTP keytab for this node. keytab: jupyterhub: @@ -33,23 +43,30 @@ gateway: apiUrl: custom: + # The full authenticator class name. class: + # Configuration fields to set on the authenticator class. config: {} livenessProbe: + # Enables the livenessProbe. enabled: true + # Configures the livenessProbe. initialDelaySeconds: 5 timeoutSeconds: 2 periodSeconds: 10 failureThreshold: 6 readinessProbe: + # Enables the readinessProbe. enabled: true + # Configures the readinessProbe. initialDelaySeconds: 5 timeoutSeconds: 2 periodSeconds: 10 failureThreshold: 3 + # nodeSelector, affinity, and tolerations the for the `api` pod conda-store-server nodeSelector: {} affinity: {} tolerations: [] @@ -58,12 +75,14 @@ gateway: backend: image: + # The image to use for both schedulers and workers name: quansight/conda-store tag: "set-by-chartpress" pullPolicy: IfNotPresent namespace: + # A mapping of environment variables to set for both schedulers and workers. environment: {} scheduler: @@ -71,10 +90,12 @@ gateway: extraContainerConfig: {} + # Cores request/limit for the scheduler. cores: request: limit: + # Memory request/limit for the scheduler. memory: request: limit: @@ -84,10 +105,12 @@ gateway: extraContainerConfig: {} + # Cores request/limit for each worker. cores: request: limit: + # Memory request/limit for each worker. memory: request: limit: @@ -97,54 +120,78 @@ gateway: controller: enabled: true + # Any annotations to add to the controller pod annotations: {} + # Resource requests/limits for the controller pod resources: {} + # Image pull secrets for controller pod imagePullSecrets: [] + # The controller log level loglevel: INFO + # Max time (in seconds) to keep around records of completed clusters. + # Default is 24 hours. completedClusterMaxAge: 86400 + # Time (in seconds) between cleanup tasks removing records of completed + # clusters. Default is 5 minutes. completedClusterCleanupPeriod: 600 + # Base delay (in seconds) for backoff when retrying after failures. backoffBaseDelay: 0.1 + # Max delay (in seconds) for backoff when retrying after failures. backoffMaxDelay: 300 + # Limit on the average number of k8s api calls per second. k8sApiRateLimit: 50 + # Limit on the maximum number of k8s api calls per second. k8sApiRateLimitBurst: 100 + # The image to use for the controller pod. image: name: quansight/conda-store-server tag: "set-by-chartpress" pullPolicy: IfNotPresent + # Settings for nodeSelector, affinity, and tolerations for the controller pods nodeSelector: {} affinity: {} tolerations: [] +# traefik nested config relates to the traefik Pod and Traefik running within it +# that is acting as a proxy for traffic towards the gateway traefik: + # Number of instances of the proxy to run replicas: 1 + # Any annotations to add to the proxy pods annotations: {} + # Resource requests/limits for the proxy pods resources: {} + # The image to use for the proxy pod image: name: traefik tag: "2.6.3" pullPolicy: IfNotPresent imagePullSecrets: [] + # Any additional arguments to forward to traefik additionalArguments: [] + # The proxy log level loglevel: WARN + # Whether to expose the dashboard on port 9000 (enable for debugging only!) dashboard: false + # Additional configuration for the traefik service service: type: LoadBalancer annotations: {} @@ -161,9 +208,14 @@ traefik: affinity: {} tolerations: [] +# rbac nested configuration relates to the choice of creating or replacing +# resources like (Cluster)Role, (Cluster)RoleBinding, and ServiceAccount. rbac: enabled: true + # Existing names to use if ClusterRoles, ClusterRoleBindings, and + # ServiceAccounts have already been created by other means (leave set to + # `null` to create all required roles at install time) controller: serviceAccountName: @@ -173,4 +225,8 @@ rbac: traefik: serviceAccountName: +# global nested configuration is accessible by all Helm charts that may depend +# on each other, but not used by this Helm chart. An entry is created here to +# validate its use and catch YAML typos via this configurations associated JSON +# schema. global: {}