Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

install questions #415

Closed
marsty339 opened this issue Jun 10, 2022 · 8 comments
Closed

install questions #415

marsty339 opened this issue Jun 10, 2022 · 8 comments
Assignees
Labels
needs-more-info Waiting for user feedback question Further information is requested stale Stale issues

Comments

@marsty339
Copy link

marsty339 commented Jun 10, 2022

when I run helm install --wait --timeout 10m tobs timescale/tobs
Error: no Secret with the name "tobs-certificate" found

helm list
NAME	NAMESPACE	REVISION	UPDATED                                	STATUS	CHART      	APP VERSION
tobs	default  	1       	2022-06-10 14:51:24.457333106 +0800 CST	failed	tobs-0.10.1	0.10.1
the helm release status is failed,when I deal with same pod not running, not something to continu..

NAME                                                        READY   STATUS      RESTARTS   AGE
alertmanager-tobs-kube-prometheus-alertmanager-0            2/2     Running     0          18m
opentelemetry-operator-controller-manager-7467bdc8b-rjg9p   2/2     Running     0          18m
prometheus-tobs-kube-prometheus-prometheus-0                2/2     Running     0          18m
tobs-connection-secret-mdwqp                                0/1     Completed   0          18m
tobs-grafana-6894864595-7844f                               3/3     Running     3          18m
tobs-grafana-db-j7rcc                                       0/1     Completed   2          18m
tobs-kube-prometheus-operator-5874b95798-55cl4              1/1     Running     0          18m
tobs-kube-state-metrics-56c568fdcc-8mdzq                    1/1     Running     0          18m
tobs-prometheus-node-exporter-fqc4j                         1/1     Running     0          18m
tobs-promlens-7f778cc958-vzdpw                              1/1     Running     0          18m
tobs-promscale-648d58d4f5-kgl52                             1/1     Running     3          18m
tobs-timescaledb-0                                          1/1     Running     0          18m

not opentelemetrycollectors cr , how can i to do?
when I delete helm release tobs, the pod tobs-opentelemetry-collector-5f5889956d-8nbb2 is creating....

@marsty339 marsty339 added the question Further information is requested label Jun 10, 2022
@paulfantom
Copy link
Contributor

We are looking into issue with no Secret with the name "tobs-certificate" found during installation.

As for:

when I delete helm release tobs, the pod tobs-opentelemetry-collector-5f5889956d-8nbb2 is creating....

We are tracking this issue in #365

@paulfantom
Copy link
Contributor

@marsty339 could you check if this is still the case when using tobs 0.11.1?

@marsty339
Copy link
Author

marsty339 commented Jun 24, 2022

 helm install --wait --timeout=10m tobs timescale/tobs
Error: secrets "tobs-pgbackrest" already exists
kubectl  get po
NAME                                                         READY   STATUS      RESTARTS      AGE
alertmanager-tobs-kube-prometheus-alertmanager-0             2/2     Running     0             45s
alertmanager-tobs-kube-prometheus-alertmanager-1             2/2     Running     0             44s
alertmanager-tobs-kube-prometheus-alertmanager-2             2/2     Running     0             43s
opentelemetry-operator-controller-manager-7c985ff6f9-l8mnr   2/2     Running     0             55s
prometheus-tobs-kube-prometheus-prometheus-0                 2/2     Running     0             45s
prometheus-tobs-kube-prometheus-prometheus-1                 2/2     Running     0             44s
tobs-connection-secret-b9k4b                                 0/1     Completed   0             55s
tobs-grafana-7cd8df4694-qg7kz                                3/3     Running     0             55s
tobs-kube-prometheus-operator-56c9986ff4-8wx6c               1/1     Running     0             55s
tobs-kube-state-metrics-677c9b4fd4-qvf7f                     1/1     Running     0             55s
tobs-prometheus-node-exporter-6wxgr                          1/1     Running     0             55s
tobs-promscale-f74b845d8-cdknf                               1/1     Running     2 (42s ago)   55s
tobs-timescaledb-0                                           1/1     Running     0             55s
The installation still has errors,And it feels very complicated to install.
tobs 0.11.1 Requires  k8s version 1.23+,but my cluster version is 1.21.4。。。

@paulfantom
Copy link
Contributor

We removed kubernetes version constraint in tobs 0.11.2. Sorry to ask you again, but could you check if everything is working fine on our latest patch release - 0.11.3?

@marsty339
Copy link
Author

marsty339 commented Jun 30, 2022

@paulfantom

helm install --wait --timeout 10m tobs timescale/tobs
Error: secrets "tobs-certificate" already exists
helm list
NAME	NAMESPACE	REVISION	UPDATED                                	STATUS	CHART      	APP VERSION
tobs	default  	1       	2022-06-30 23:56:44.229760508 +0800 CST	failed	tobs-0.11.3	0.11.3
kubectl  get po
NAME                                                         READY   STATUS      RESTARTS       AGE
alertmanager-tobs-kube-prometheus-alertmanager-0             2/2     Running     0              2m17s
alertmanager-tobs-kube-prometheus-alertmanager-1             2/2     Running     0              2m17s
alertmanager-tobs-kube-prometheus-alertmanager-2             2/2     Running     0              2m17s
opentelemetry-operator-controller-manager-7c985ff6f9-2hnw2   2/2     Running     0              2m31s
prometheus-tobs-kube-prometheus-prometheus-0                 2/2     Running     0              2m16s
prometheus-tobs-kube-prometheus-prometheus-1                 2/2     Running     0              2m16s
tobs-connection-secret-ng466                                 0/1     Completed   0              2m31s
tobs-grafana-5dfc6fb499-4vkk5                                3/3     Running     0              2m31s
tobs-grafana-db-9vp8h                                        0/1     Completed   2              2m31s
tobs-kube-prometheus-operator-56c9986ff4-4rz5p               1/1     Running     0              2m31s
tobs-kube-state-metrics-677c9b4fd4-ljh9g                     1/1     Running     0              2m31s
tobs-prometheus-node-exporter-fnfh9                          1/1     Running     0              2m31s
tobs-promscale-f74b845d8-jwf4w                               1/1     Running     3 (113s ago)   2m31s
tobs-timescaledb-0                                           1/1     Running     0              2m31s

@paulfantom
Copy link
Contributor

paulfantom commented Jul 4, 2022

I cannot reproduce this.

When doing the following:

helm repo add timescale https://charts.timescale.com/
helm repo update
helm install --wait --timeout 10m test timescale/tobs

I get the following, correct, output from the installation:

W0704 11:55:15.004749  673465 warnings.go:70] spec.template.spec.containers[0].env[2].name: duplicate name "TOBS_TELEMETRY_INSTALLED_BY"
W0704 11:55:15.004768  673465 warnings.go:70] spec.template.spec.containers[0].env[3].name: duplicate name "TOBS_TELEMETRY_VERSION"
NAME: test
LAST DEPLOYED: Mon Jul  4 11:54:55 2022
NAMESPACE: default
STATUS: deployed
REVISION: 1
NOTES:
#######################################################################################################################
👋🏽 Welcome to tobs, The Observability Stack for Kubernetes

✨ Auto-configured and deployed:
🔥 Kube-Prometheus
🐯 In-cluster TimescaleDB
🤝 Promscale
📈 Grafana
🚀 OpenTelemetry

#######################################################################################################################
🔥 PROMETHEUS NOTES:
#######################################################################################################################

Prometheus can be accessed via port 9090 on the following DNS name from within your cluster:
    tobs-kube-prometheus-prometheus.default.svc

Get the Prometheus server URL by running these commands in the same shell:
    kubectl --namespace default port-forward service/tobs-kube-prometheus-prometheus 9090:9090


#######################################################################################################################
🔥 ALERTMANAGER NOTES:
#######################################################################################################################

The Alertmanager can be accessed via port 9093 on the following DNS name
from within your cluster:
    tobs-kube-prometheus-alertmanager.default.svc

Get the Alertmanager URL by running these commands in the same shell:
    kubectl --namespace default port-forward service/tobs-kube-prometheus-alertmanager 9093:9093

WARNING! Persistence is disabled on AlertManager. You will lose your data when
         the AlertManager pod is terminated.

#######################################################################################################################
🐯 TIMESCALEDB NOTES:
#######################################################################################################################

TimescaleDB can be accessed via port 5432 on the following DNS name
from within your cluster:
    test.default.svc

To get your password for superuser run:
    # superuser password
    PGPASSWORD_POSTGRES=$(
      kubectl get secret --namespace default \
        test-credentials \
        -o jsonpath="{.data.PATRONI_SUPERUSER_PASSWORD}" |\
      base64 --decode \
    )
    echo $PGPASSWORD_POSTGRES

    # admin password
    PGPASSWORD_ADMIN=$(\
      kubectl get secret --namespace default \
        test-credentials \
        -o jsonpath="{.data.PATRONI_admin_PASSWORD}" |\
      base64 --decode \
    )
    echo $PGPASSWORD_ADMIN

To connect to your database, chose one of these options:

1. Run a postgres pod and connect using the psql cli:
    # login as superuser
    kubectl run -it --rm psql --image=postgres --env "PGPASSWORD=$PGPASSWORD_POSTGRES" --command --\
      psql -U postgres -h test.default.svc postgres

    # login as admin
    kubectl run -it --rm psql --image=postgres --env "PGPASSWORD=$PGPASSWORD_ADMIN" --command --\
      psql -U admin -h test.default.svc postgres

2. Directly execute a psql session on the master node
    MASTER_POD=$(\
      kubectl get pod -o name --namespace default -l release=test,role=master \
    )
    kubectl exec -it --namespace default ${MASTERPOD} -- psql -U postgres

#######################################################################################################################
🚀  OPENTELEMETRY NOTES:
#######################################################################################################################

The OpenTelemetry collector is deployed to collect traces.

OpenTelemetry collector can be accessed with the following DNS name from within your cluster:
    test-opentelemetry-collector.default.svc

#######################################################################################################################
📈 GRAFANA NOTES:
#######################################################################################################################

The Grafana server can be accessed via port 80 on the following DNS name from within your cluster:
   test-grafana.default.svc

You can access grafana locally by executing:
    kubectl --namespace default port-forward service/test-grafana 8080:80

   Then you can point your browser to http://127.0.0.1:8080/.

Grafana persistence is enabled, and you did an upgrade. If you don't have the password
for 'admin', it can not be retrieved again, you need to reset it (see next paragraph).

To reset the admin user password you can use grafana-cli from inside the pod by executing:
    GRAFANA_POD="$(kubectl get pod -o name --namespace default -l app.kubernetes.io/name=grafana)"
    kubectl exec -it ${GRAFANA_POD} -c grafana -- grafana-cli admin reset-admin-password <password-you-want-to-set>

🚀 Happy observing!

If possible, I would recommend removing namespace in which tobs should be installed and start from scratch. However if this is not possible, I recommend to manually remove tobs-certificate secret and try again.

Pro tip: Use helm install --atomic to instruct helm to remove objects when deployment failed. This way you shouldn't run into issues with stale kubernetes objects when trying out things.

@paulfantom paulfantom self-assigned this Jul 4, 2022
@paulfantom paulfantom added the needs-more-info Waiting for user feedback label Jul 4, 2022
@github-actions
Copy link

github-actions bot commented Aug 4, 2022

This issue went stale because it was not updated in a month. Please consider updating it to improve the quality of the project.

@github-actions github-actions bot added the stale Stale issues label Aug 4, 2022
@github-actions
Copy link

github-actions bot commented Sep 5, 2022

This issue was closed because it has been stalled for 30 days with no activity.

@github-actions github-actions bot closed this as completed Sep 5, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
needs-more-info Waiting for user feedback question Further information is requested stale Stale issues
Projects
None yet
Development

No branches or pull requests

2 participants