-
Notifications
You must be signed in to change notification settings - Fork 4.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add citadel metricset for Istio Metricbeat module #15990
Conversation
Signed-off-by: chrismark <[email protected]>
description: > | ||
The type of the respective grpc service | ||
|
||
- name: secret_controller_svc_acc_created_cert_count |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This must be secret_controller_svc_acc_created_cert.count
to follow naming conventions
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for mentioning. Will change it.
type: long | ||
description: > | ||
The number of certificates created due to service account creation. | ||
- name: server_root_cert_expiry_timestamp |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'd call it server_root_cert_expiry_seconds
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for suggesting. Will change it.
description: > | ||
Total number of RPCs started on the server. | ||
|
||
- name: grpc.server.handling.latency.ms.bucket.* |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
2 things here:
- Usually the unit comes at the end of the field, it's tricky in this case because of the wildcard anyways:
grpc.server.handling.latency.bucket.ms.*
- I wonder if we should store this operations instead of "doing" them in Kibana.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is common pattern so far for this kind of Metrics (Prometheus histograms). See
- name: duration.us.bucket.* |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great, thank you for explaining :)
object_type: long | ||
description: > | ||
The response latency (milliseconds) of gRPC that had been application-level handled by the server. | ||
- name: grpc.server.handling.latency.ms.sum |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Unit at the end :) grpc.server.handling.latency.sum.ms
, same thing with the sum operation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
same as above this is a prometheus spesific concern.
format: duration | ||
description: > | ||
The response latency of gRPC, sum of latencies in milliseconds | ||
- name: grpc.server.handling.latency.ms.count |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Unit at the end
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
same as above this is a prometheus spesific concern.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Left some comments. More like questions than requests. Maybe none of them apply. I suggest to create a quick dashboard with the data coming from the metricset to see how much you can visualize them. It's a good exercise when creating a metricset.
Thank you for reviewing @sayden ! Will push the adjustments soonish. Regarding the Prometheus Histogram types as I mentioned inline, this is something that we follow already for this kind of metrics. However this might be a subject of refactoring anyway in the near feature while we are working on #14843. In this case I would suggest we proceed with this currently and have all of them to be refactored with the new type introduction. wdyt? Regarding the Dashboard, it is planned already and listed in the "TODOs" #15505 (comment). |
Yeap, yeap, let's proceed. I just wanted to make sure that we weren't missing anything :) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Signed-off-by: chrismark <[email protected]>
(cherry picked from commit e9b0be3)
This PR adds citadel metricset for istio Module.
The metrics will be collected from the respective Prometheus exporter endpoint:
istio-citadel.istio-system:15014: The istio-citadel job returns all Citadel-generated metrics.
Part of: #15505
Manual testing
kubectl -n istio-system port-forward svc/istio-citadel 15014:15014
istio
module, configure it in metricbeat.yaml and start monitoring ISTIOcitadel